var/home/core/zuul-output/0000755000175000017500000000000015110664227014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110677251015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005053251515110677242017710 0ustar rootrootNov 23 20:08:10 crc systemd[1]: Starting Kubernetes Kubelet... Nov 23 20:08:10 crc restorecon[4644]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:10 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 23 20:08:11 crc restorecon[4644]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 23 20:08:12 crc kubenswrapper[4726]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 20:08:12 crc kubenswrapper[4726]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 23 20:08:12 crc kubenswrapper[4726]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 20:08:12 crc kubenswrapper[4726]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 20:08:12 crc kubenswrapper[4726]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 23 20:08:12 crc kubenswrapper[4726]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.299388 4726 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317704 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317747 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317758 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317767 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317776 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317785 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317794 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317802 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317811 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317820 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317828 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317836 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317845 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317853 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317861 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317905 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317913 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317921 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317928 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317937 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317945 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317953 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317960 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317969 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317977 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317984 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.317995 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318006 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318014 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318024 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318033 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318054 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318063 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318071 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318079 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318087 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318094 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318102 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318110 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318121 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318133 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318142 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318152 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318160 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318168 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318179 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318190 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318199 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318207 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318216 4726 feature_gate.go:330] unrecognized feature gate: Example Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318225 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318236 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318245 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318254 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318263 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318272 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318283 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318293 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318303 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318312 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318320 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318330 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318342 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318353 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318363 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318371 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318381 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318390 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318398 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318406 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.318414 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319537 4726 flags.go:64] FLAG: --address="0.0.0.0" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319571 4726 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319591 4726 flags.go:64] FLAG: --anonymous-auth="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319603 4726 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319615 4726 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319624 4726 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319638 4726 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319652 4726 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319664 4726 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319678 4726 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319689 4726 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319713 4726 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319723 4726 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319733 4726 flags.go:64] FLAG: --cgroup-root="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319742 4726 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319753 4726 flags.go:64] FLAG: --client-ca-file="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319764 4726 flags.go:64] FLAG: --cloud-config="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319777 4726 flags.go:64] FLAG: --cloud-provider="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319788 4726 flags.go:64] FLAG: --cluster-dns="[]" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319808 4726 flags.go:64] FLAG: --cluster-domain="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319817 4726 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319827 4726 flags.go:64] FLAG: --config-dir="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319835 4726 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319845 4726 flags.go:64] FLAG: --container-log-max-files="5" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319857 4726 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319898 4726 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319909 4726 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319920 4726 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319929 4726 flags.go:64] FLAG: --contention-profiling="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319939 4726 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319951 4726 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319963 4726 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319975 4726 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.319989 4726 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320001 4726 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320012 4726 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320024 4726 flags.go:64] FLAG: --enable-load-reader="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320036 4726 flags.go:64] FLAG: --enable-server="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320048 4726 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320070 4726 flags.go:64] FLAG: --event-burst="100" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320080 4726 flags.go:64] FLAG: --event-qps="50" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320090 4726 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320099 4726 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320108 4726 flags.go:64] FLAG: --eviction-hard="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320119 4726 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320128 4726 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320137 4726 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320160 4726 flags.go:64] FLAG: --eviction-soft="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320169 4726 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320178 4726 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320186 4726 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320195 4726 flags.go:64] FLAG: --experimental-mounter-path="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320204 4726 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320213 4726 flags.go:64] FLAG: --fail-swap-on="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320222 4726 flags.go:64] FLAG: --feature-gates="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320233 4726 flags.go:64] FLAG: --file-check-frequency="20s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320242 4726 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320251 4726 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320261 4726 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320270 4726 flags.go:64] FLAG: --healthz-port="10248" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320279 4726 flags.go:64] FLAG: --help="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320288 4726 flags.go:64] FLAG: --hostname-override="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320297 4726 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320307 4726 flags.go:64] FLAG: --http-check-frequency="20s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320317 4726 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320326 4726 flags.go:64] FLAG: --image-credential-provider-config="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320335 4726 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320344 4726 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320354 4726 flags.go:64] FLAG: --image-service-endpoint="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320362 4726 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320371 4726 flags.go:64] FLAG: --kube-api-burst="100" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320380 4726 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320389 4726 flags.go:64] FLAG: --kube-api-qps="50" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320399 4726 flags.go:64] FLAG: --kube-reserved="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320408 4726 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320417 4726 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320426 4726 flags.go:64] FLAG: --kubelet-cgroups="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320435 4726 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320444 4726 flags.go:64] FLAG: --lock-file="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320452 4726 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320462 4726 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320471 4726 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320484 4726 flags.go:64] FLAG: --log-json-split-stream="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320504 4726 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320514 4726 flags.go:64] FLAG: --log-text-split-stream="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320523 4726 flags.go:64] FLAG: --logging-format="text" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320531 4726 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320541 4726 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320550 4726 flags.go:64] FLAG: --manifest-url="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320559 4726 flags.go:64] FLAG: --manifest-url-header="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320571 4726 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320580 4726 flags.go:64] FLAG: --max-open-files="1000000" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320590 4726 flags.go:64] FLAG: --max-pods="110" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320599 4726 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320610 4726 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320622 4726 flags.go:64] FLAG: --memory-manager-policy="None" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320633 4726 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320645 4726 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320657 4726 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320669 4726 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320696 4726 flags.go:64] FLAG: --node-status-max-images="50" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320705 4726 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320714 4726 flags.go:64] FLAG: --oom-score-adj="-999" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320723 4726 flags.go:64] FLAG: --pod-cidr="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320732 4726 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320745 4726 flags.go:64] FLAG: --pod-manifest-path="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320754 4726 flags.go:64] FLAG: --pod-max-pids="-1" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320763 4726 flags.go:64] FLAG: --pods-per-core="0" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320772 4726 flags.go:64] FLAG: --port="10250" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320782 4726 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320792 4726 flags.go:64] FLAG: --provider-id="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320800 4726 flags.go:64] FLAG: --qos-reserved="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320809 4726 flags.go:64] FLAG: --read-only-port="10255" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320818 4726 flags.go:64] FLAG: --register-node="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320827 4726 flags.go:64] FLAG: --register-schedulable="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320836 4726 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320852 4726 flags.go:64] FLAG: --registry-burst="10" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320861 4726 flags.go:64] FLAG: --registry-qps="5" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320896 4726 flags.go:64] FLAG: --reserved-cpus="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320929 4726 flags.go:64] FLAG: --reserved-memory="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320941 4726 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320950 4726 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320959 4726 flags.go:64] FLAG: --rotate-certificates="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320968 4726 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320977 4726 flags.go:64] FLAG: --runonce="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320986 4726 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.320995 4726 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321004 4726 flags.go:64] FLAG: --seccomp-default="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321014 4726 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321022 4726 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321033 4726 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321042 4726 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321051 4726 flags.go:64] FLAG: --storage-driver-password="root" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321059 4726 flags.go:64] FLAG: --storage-driver-secure="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321069 4726 flags.go:64] FLAG: --storage-driver-table="stats" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321077 4726 flags.go:64] FLAG: --storage-driver-user="root" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321088 4726 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321099 4726 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321109 4726 flags.go:64] FLAG: --system-cgroups="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321120 4726 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321137 4726 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321150 4726 flags.go:64] FLAG: --tls-cert-file="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321161 4726 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321188 4726 flags.go:64] FLAG: --tls-min-version="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321200 4726 flags.go:64] FLAG: --tls-private-key-file="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321211 4726 flags.go:64] FLAG: --topology-manager-policy="none" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321222 4726 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321231 4726 flags.go:64] FLAG: --topology-manager-scope="container" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321240 4726 flags.go:64] FLAG: --v="2" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321252 4726 flags.go:64] FLAG: --version="false" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321267 4726 flags.go:64] FLAG: --vmodule="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321294 4726 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.321304 4726 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321537 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321583 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321599 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321611 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321622 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321633 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321644 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321662 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321672 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321682 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321692 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321702 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321711 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321718 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321727 4726 feature_gate.go:330] unrecognized feature gate: Example Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321734 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321742 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321749 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321757 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321766 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321774 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321781 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321789 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321796 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321804 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321813 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321820 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321828 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321835 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321843 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321851 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321858 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321900 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321910 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321920 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321930 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321940 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321954 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321976 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321987 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.321996 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322004 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322013 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322020 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322029 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322037 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322045 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322053 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322060 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322070 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322080 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322089 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322097 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322107 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322116 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322124 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322133 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322141 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322149 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322157 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322167 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322177 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322185 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322193 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322202 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322211 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322219 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322227 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322234 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322242 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.322252 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.322271 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.337338 4726 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.337385 4726 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337546 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337604 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337618 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337629 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337640 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337649 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337658 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337668 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337678 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337688 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337697 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337707 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337718 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337729 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337739 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337749 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337759 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337769 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337779 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337788 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337798 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337807 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337821 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337833 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337844 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337854 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337864 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337913 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337925 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337935 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337943 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337951 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337960 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337969 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337981 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.337992 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338000 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338010 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338021 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338030 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338039 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338048 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338056 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338064 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338072 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338081 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338088 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338097 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338104 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338112 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338120 4726 feature_gate.go:330] unrecognized feature gate: Example Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338128 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338136 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338145 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338153 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338161 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338169 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338177 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338184 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338192 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338200 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338208 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338215 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338223 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338233 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338242 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338250 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338258 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338266 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338274 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338285 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.338297 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338514 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338529 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338538 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338546 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338555 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338563 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338572 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338580 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338588 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338596 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338604 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338612 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338620 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338628 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338635 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338645 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338652 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338660 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338670 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338678 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338686 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338694 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338702 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338711 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338719 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338727 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338734 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338742 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338750 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338758 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338766 4726 feature_gate.go:330] unrecognized feature gate: Example Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338773 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338781 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338789 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338798 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338806 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338814 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338821 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338829 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338837 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338844 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338852 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338860 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338903 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338916 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338926 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338935 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338943 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338952 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338960 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338967 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338975 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338983 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338990 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.338998 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339006 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339013 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339021 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339029 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339037 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339047 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339056 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339067 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339077 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339087 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339097 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339107 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339115 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339124 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339134 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.339146 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.339157 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.339401 4726 server.go:940] "Client rotation is on, will bootstrap in background" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.345344 4726 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.345480 4726 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.348014 4726 server.go:997] "Starting client certificate rotation" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.348065 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.349059 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-29 10:35:10.893373433 +0000 UTC Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.349184 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 854h26m58.544194444s for next certificate rotation Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.378212 4726 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.380973 4726 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.401974 4726 log.go:25] "Validated CRI v1 runtime API" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.437496 4726 log.go:25] "Validated CRI v1 image API" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.439748 4726 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.446779 4726 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-23-20-01-59-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.446842 4726 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.474078 4726 manager.go:217] Machine: {Timestamp:2025-11-23 20:08:12.471603824 +0000 UTC m=+0.620644860 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:d521adaa-c97b-4677-a80e-46a3d7f7a33d BootID:998bc227-8b31-4e93-8b90-99ca7f2d4cd9 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:45:d8:85 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:45:d8:85 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e4:82:f0 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:12:74:51 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:76:fb:dd Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:78:95:b6 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:4a:b2:ad Speed:-1 Mtu:1496} {Name:eth10 MacAddress:f2:55:bf:98:f4:d7 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5a:3e:6e:b1:6c:dc Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.474526 4726 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.474786 4726 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.477995 4726 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.478365 4726 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.478423 4726 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.478736 4726 topology_manager.go:138] "Creating topology manager with none policy" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.478757 4726 container_manager_linux.go:303] "Creating device plugin manager" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.479451 4726 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.479497 4726 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.479931 4726 state_mem.go:36] "Initialized new in-memory state store" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.480502 4726 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.487309 4726 kubelet.go:418] "Attempting to sync node with API server" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.487359 4726 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.487418 4726 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.487453 4726 kubelet.go:324] "Adding apiserver pod source" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.487476 4726 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.495097 4726 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.496071 4726 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.496543 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.496545 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.496797 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.496830 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.498916 4726 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500467 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500509 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500524 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500538 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500559 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500573 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500586 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500612 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500632 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500652 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500677 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500695 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.500762 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.501399 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.501524 4726 server.go:1280] "Started kubelet" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.501749 4726 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.502086 4726 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.504154 4726 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.505157 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.505179 4726 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 23 20:08:12 crc systemd[1]: Started Kubernetes Kubelet. Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.505563 4726 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.505624 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 13:39:14.793774131 +0000 UTC Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.505737 4726 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.505758 4726 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.505909 4726 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.505977 4726 server.go:460] "Adding debug handlers to kubelet server" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.507363 4726 factory.go:55] Registering systemd factory Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.507395 4726 factory.go:221] Registration of the systemd container factory successfully Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.507808 4726 factory.go:153] Registering CRI-O factory Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.507836 4726 factory.go:221] Registration of the crio container factory successfully Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.507954 4726 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.507991 4726 factory.go:103] Registering Raw factory Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.508016 4726 manager.go:1196] Started watching for new ooms in manager Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.508663 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="200ms" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.508841 4726 manager.go:319] Starting recovery of all containers Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.508854 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.508985 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.509213 4726 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.110:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187abb9e44bd9bf4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-23 20:08:12.501474292 +0000 UTC m=+0.650515308,LastTimestamp:2025-11-23 20:08:12.501474292 +0000 UTC m=+0.650515308,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527691 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527753 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527772 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527792 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527809 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527824 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527839 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527855 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527900 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527920 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527938 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527956 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527968 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527985 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.527998 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528015 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528030 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528048 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528062 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528076 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528092 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528107 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528122 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528138 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528156 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528170 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528188 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528203 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528218 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528232 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528246 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528263 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528277 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528292 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528306 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528321 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528338 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528355 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528379 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528397 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528413 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528431 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528448 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528466 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528483 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528499 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528518 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528536 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528551 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528568 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528583 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528599 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528622 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528639 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528655 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528671 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528687 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528701 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528718 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528733 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528747 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528762 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528780 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528798 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528813 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528846 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528864 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528901 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528916 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528931 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528948 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528968 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.528987 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529005 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529022 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529039 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529056 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529072 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529089 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529107 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529128 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529180 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529198 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529214 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529234 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529250 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529270 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529287 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529304 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529323 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529341 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529358 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529375 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529404 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529421 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529441 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529456 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529473 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529489 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529504 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529520 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529536 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529551 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529566 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529590 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529608 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529627 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529653 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529696 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529714 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529738 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529758 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529776 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529793 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529809 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529826 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529842 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529857 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.529897 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532436 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532462 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532481 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532500 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532520 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532539 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532556 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532572 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.532592 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534423 4726 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534473 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534495 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534513 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534534 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534553 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534571 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534588 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534611 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534628 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534645 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534662 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534681 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534697 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534714 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534734 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534751 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534766 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534785 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534801 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534833 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534850 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534889 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534908 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534925 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534940 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534958 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534976 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.534993 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535009 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535024 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535041 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535058 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535077 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535096 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535112 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535127 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535144 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535161 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535176 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535192 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535209 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535226 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535245 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535276 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535296 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535313 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535332 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535349 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535367 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535407 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535427 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535443 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535460 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535477 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535494 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535510 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535525 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535541 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535556 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535571 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535586 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535603 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535621 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535637 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535653 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535672 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535690 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535706 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535721 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535740 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535756 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535775 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535793 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535809 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535825 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535843 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535859 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535897 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535917 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535933 4726 reconstruct.go:97] "Volume reconstruction finished" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.535944 4726 reconciler.go:26] "Reconciler: start to sync state" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.545188 4726 manager.go:324] Recovery completed Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.560833 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.562375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.562410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.562421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.563326 4726 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.563345 4726 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.563364 4726 state_mem.go:36] "Initialized new in-memory state store" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.585047 4726 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.587384 4726 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.587732 4726 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.587797 4726 kubelet.go:2335] "Starting kubelet main sync loop" Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.587966 4726 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.590058 4726 policy_none.go:49] "None policy: Start" Nov 23 20:08:12 crc kubenswrapper[4726]: W1123 20:08:12.590497 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.590587 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.591036 4726 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.591081 4726 state_mem.go:35] "Initializing new in-memory state store" Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.608490 4726 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.642444 4726 manager.go:334] "Starting Device Plugin manager" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.642504 4726 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.642522 4726 server.go:79] "Starting device plugin registration server" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.643040 4726 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.643080 4726 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.643397 4726 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.643505 4726 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.643515 4726 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.660895 4726 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.688143 4726 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.688312 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.689712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.689783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.689797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.689980 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.690145 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.690200 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691073 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691129 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691147 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691352 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691455 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691501 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.691550 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692563 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692738 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692746 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692831 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692907 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.692955 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694024 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694067 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694106 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694123 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694083 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694316 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694476 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.694522 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.695440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.695473 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.695490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.695518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.695547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.695566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.695785 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.695836 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.698732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.698780 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.698799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.709276 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="400ms" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738623 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738657 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738677 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738693 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738710 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738728 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738746 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738786 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738808 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738914 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738966 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.738999 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.739044 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.739079 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.739107 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.744038 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.745369 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.745410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.745425 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.745453 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.745943 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.840780 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841176 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841226 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841295 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841356 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841371 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841388 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841430 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841447 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841477 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841485 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841532 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841541 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841578 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841628 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841627 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841689 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841726 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841746 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841779 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841810 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841829 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841837 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841964 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.841991 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.842025 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.842071 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.842112 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.842150 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.842188 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.946079 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.947803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.947893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.947932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:12 crc kubenswrapper[4726]: I1123 20:08:12.947975 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 20:08:12 crc kubenswrapper[4726]: E1123 20:08:12.948556 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.015199 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.021815 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.045252 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:13 crc kubenswrapper[4726]: W1123 20:08:13.064153 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-47473eaaa887063fd845f12d2a5ae8524451747b19a12c7b4ff76eef4e41a82a WatchSource:0}: Error finding container 47473eaaa887063fd845f12d2a5ae8524451747b19a12c7b4ff76eef4e41a82a: Status 404 returned error can't find the container with id 47473eaaa887063fd845f12d2a5ae8524451747b19a12c7b4ff76eef4e41a82a Nov 23 20:08:13 crc kubenswrapper[4726]: W1123 20:08:13.064500 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-66bb778f00d5ab938b0dbe9224a08a4827df87446f37ff8677013fbba976a71e WatchSource:0}: Error finding container 66bb778f00d5ab938b0dbe9224a08a4827df87446f37ff8677013fbba976a71e: Status 404 returned error can't find the container with id 66bb778f00d5ab938b0dbe9224a08a4827df87446f37ff8677013fbba976a71e Nov 23 20:08:13 crc kubenswrapper[4726]: W1123 20:08:13.071307 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-89643525282a55f4416092c46ebb1705aef60461524902e9cfb9a032e53b3e1c WatchSource:0}: Error finding container 89643525282a55f4416092c46ebb1705aef60461524902e9cfb9a032e53b3e1c: Status 404 returned error can't find the container with id 89643525282a55f4416092c46ebb1705aef60461524902e9cfb9a032e53b3e1c Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.074790 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.081096 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:13 crc kubenswrapper[4726]: W1123 20:08:13.108595 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a6565d0daa89c627be11e412f7878a56da16d171ebb7de9684aa9ee49db3231d WatchSource:0}: Error finding container a6565d0daa89c627be11e412f7878a56da16d171ebb7de9684aa9ee49db3231d: Status 404 returned error can't find the container with id a6565d0daa89c627be11e412f7878a56da16d171ebb7de9684aa9ee49db3231d Nov 23 20:08:13 crc kubenswrapper[4726]: E1123 20:08:13.109793 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="800ms" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.349553 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.358632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.358677 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.358686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.358715 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 20:08:13 crc kubenswrapper[4726]: E1123 20:08:13.359182 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 23 20:08:13 crc kubenswrapper[4726]: W1123 20:08:13.463288 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:13 crc kubenswrapper[4726]: E1123 20:08:13.463410 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.502907 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.506042 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 20:48:00.289350143 +0000 UTC Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.595308 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a6565d0daa89c627be11e412f7878a56da16d171ebb7de9684aa9ee49db3231d"} Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.599785 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"18cc380d9da035aab8b64df9209466a1a34415ee159190e9c2a950e1a6ddf5cf"} Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.601940 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"89643525282a55f4416092c46ebb1705aef60461524902e9cfb9a032e53b3e1c"} Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.603515 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"66bb778f00d5ab938b0dbe9224a08a4827df87446f37ff8677013fbba976a71e"} Nov 23 20:08:13 crc kubenswrapper[4726]: I1123 20:08:13.604656 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"47473eaaa887063fd845f12d2a5ae8524451747b19a12c7b4ff76eef4e41a82a"} Nov 23 20:08:13 crc kubenswrapper[4726]: W1123 20:08:13.620190 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:13 crc kubenswrapper[4726]: E1123 20:08:13.620388 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:13 crc kubenswrapper[4726]: W1123 20:08:13.871407 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:13 crc kubenswrapper[4726]: E1123 20:08:13.871491 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:13 crc kubenswrapper[4726]: E1123 20:08:13.910797 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="1.6s" Nov 23 20:08:14 crc kubenswrapper[4726]: W1123 20:08:14.057697 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:14 crc kubenswrapper[4726]: E1123 20:08:14.057808 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.159725 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.160908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.160941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.160953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.160978 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 20:08:14 crc kubenswrapper[4726]: E1123 20:08:14.161277 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.502516 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.506707 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 06:47:20.150455286 +0000 UTC Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.506728 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 826h39m5.643729565s for next certificate rotation Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.611593 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24"} Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.611662 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422"} Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.611685 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563"} Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.611703 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4"} Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.611602 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.613229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.613285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.613303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.615109 4726 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5" exitCode=0 Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.615247 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.615761 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5"} Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.616210 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.616231 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.616273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.617610 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.618259 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.618272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.618281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.620674 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81" exitCode=0 Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.620831 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.621158 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81"} Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.624668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.624799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.624823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.626171 4726 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4" exitCode=0 Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.626421 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4"} Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.626694 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.628753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.629045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.629209 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.631265 4726 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01" exitCode=0 Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.631337 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01"} Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.631481 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.632406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.632430 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.632440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.889894 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:14 crc kubenswrapper[4726]: I1123 20:08:14.920251 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.502320 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:15 crc kubenswrapper[4726]: E1123 20:08:15.512011 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.110:6443: connect: connection refused" interval="3.2s" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.640729 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.640784 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.640796 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.640807 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.648262 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a" exitCode=0 Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.648367 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.648374 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.652023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.652055 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.652065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.661158 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0025d40e9e0a62c038a83d3258a2f3e72c4e33423152f7ee0f9e53430e22e23b"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.661275 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.662323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.662346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.662356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.669746 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.670616 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.671031 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.671070 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.671087 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df"} Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.672247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.672283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.672295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.673141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.673229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.673249 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.762424 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.769778 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.769824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.769838 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:15 crc kubenswrapper[4726]: I1123 20:08:15.769917 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 20:08:15 crc kubenswrapper[4726]: E1123 20:08:15.770456 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.110:6443: connect: connection refused" node="crc" Nov 23 20:08:15 crc kubenswrapper[4726]: W1123 20:08:15.949944 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:15 crc kubenswrapper[4726]: E1123 20:08:15.950075 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:16 crc kubenswrapper[4726]: W1123 20:08:16.145707 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.110:6443: connect: connection refused Nov 23 20:08:16 crc kubenswrapper[4726]: E1123 20:08:16.145864 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.110:6443: connect: connection refused" logger="UnhandledError" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.679526 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3"} Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.680308 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.681782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.681915 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.681945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.682783 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f" exitCode=0 Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.682904 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.682954 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.682979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f"} Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.683017 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.683038 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.683164 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.683273 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.684136 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.684192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.684218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.684696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.684753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.684777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.685265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.685316 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.685340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.685736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.685790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:16 crc kubenswrapper[4726]: I1123 20:08:16.685814 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.692739 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404"} Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.692809 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.692811 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8"} Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.692928 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.692938 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32"} Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.692963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6"} Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.694148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.694201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.694236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.841315 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.841556 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.845988 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.846052 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.846072 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.936197 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.936618 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.938270 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.938336 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:17 crc kubenswrapper[4726]: I1123 20:08:17.938356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.704006 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1"} Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.704260 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.707438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.707508 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.707535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.971435 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.973390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.973630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.973824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:18 crc kubenswrapper[4726]: I1123 20:08:18.974065 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.008813 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.206475 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.206698 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.208485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.208535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.208552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.706408 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.707621 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.707685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:19 crc kubenswrapper[4726]: I1123 20:08:19.707708 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.615921 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.616103 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.617816 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.617912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.617934 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.651804 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.651931 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.653669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.653734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.653754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.710153 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.711657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.711704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:20 crc kubenswrapper[4726]: I1123 20:08:20.711723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.012750 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.188948 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.189462 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.191215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.191462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.191623 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.712145 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.713698 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.713812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:21 crc kubenswrapper[4726]: I1123 20:08:21.713949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:22 crc kubenswrapper[4726]: E1123 20:08:22.661189 4726 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.010536 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.010758 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.012164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.012199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.012208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.016796 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.718709 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.719616 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.719646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:24 crc kubenswrapper[4726]: I1123 20:08:24.719655 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:26 crc kubenswrapper[4726]: I1123 20:08:26.453090 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 23 20:08:26 crc kubenswrapper[4726]: I1123 20:08:26.453148 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 23 20:08:26 crc kubenswrapper[4726]: I1123 20:08:26.460002 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 23 20:08:26 crc kubenswrapper[4726]: I1123 20:08:26.460058 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 23 20:08:27 crc kubenswrapper[4726]: I1123 20:08:27.011142 4726 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 23 20:08:27 crc kubenswrapper[4726]: I1123 20:08:27.011253 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.044959 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.045112 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.046277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.046339 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.046357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.060289 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.195251 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.195539 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.197023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.197119 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.197222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.200338 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.434440 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.438015 4726 trace.go:236] Trace[1686021672]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 20:08:17.006) (total time: 14431ms): Nov 23 20:08:31 crc kubenswrapper[4726]: Trace[1686021672]: ---"Objects listed" error: 14431ms (20:08:31.437) Nov 23 20:08:31 crc kubenswrapper[4726]: Trace[1686021672]: [14.431278411s] [14.431278411s] END Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.438038 4726 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.438122 4726 trace.go:236] Trace[995918802]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 20:08:21.118) (total time: 10319ms): Nov 23 20:08:31 crc kubenswrapper[4726]: Trace[995918802]: ---"Objects listed" error: 10319ms (20:08:31.438) Nov 23 20:08:31 crc kubenswrapper[4726]: Trace[995918802]: [10.319534223s] [10.319534223s] END Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.438146 4726 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.440809 4726 trace.go:236] Trace[1243060292]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 20:08:21.019) (total time: 10421ms): Nov 23 20:08:31 crc kubenswrapper[4726]: Trace[1243060292]: ---"Objects listed" error: 10421ms (20:08:31.440) Nov 23 20:08:31 crc kubenswrapper[4726]: Trace[1243060292]: [10.421641585s] [10.421641585s] END Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.440827 4726 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.441447 4726 trace.go:236] Trace[874193018]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (23-Nov-2025 20:08:16.519) (total time: 14921ms): Nov 23 20:08:31 crc kubenswrapper[4726]: Trace[874193018]: ---"Objects listed" error: 14921ms (20:08:31.441) Nov 23 20:08:31 crc kubenswrapper[4726]: Trace[874193018]: [14.921738431s] [14.921738431s] END Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.441624 4726 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.442554 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.444119 4726 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.498272 4726 apiserver.go:52] "Watching apiserver" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.501003 4726 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.501254 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.501530 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.501669 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.501666 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.501831 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.501836 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.501911 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.501949 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.501976 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.502204 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.506316 4726 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.507379 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.507469 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.507710 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.507786 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.510428 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.521634 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.521842 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.521951 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.522026 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.522184 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.522305 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.522670 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.522997 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545135 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545420 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545482 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545502 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545521 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545536 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545552 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545570 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545584 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545599 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545614 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545629 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545643 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545658 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545672 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545687 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545703 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545717 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545732 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545747 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545760 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545777 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545797 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545813 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545829 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545846 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545862 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545895 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545913 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545929 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545945 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545963 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545979 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.545995 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546010 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546029 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546054 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546069 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546088 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546103 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546120 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546138 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546156 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546170 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546191 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546206 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546239 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546254 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546270 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546287 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546304 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546320 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546336 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546354 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546372 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546387 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546411 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546426 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546440 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546455 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546469 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546550 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546570 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546585 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546600 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546617 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546631 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546645 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546661 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546680 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546694 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546710 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546725 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546740 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546765 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546781 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546797 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546813 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546828 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546843 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546858 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546891 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546907 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546921 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546936 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546951 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546967 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546983 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.546998 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547016 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547030 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547045 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547061 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547078 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547093 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547111 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547128 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547144 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547159 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547173 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547189 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547219 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547234 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547250 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547267 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547284 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547299 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547315 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547332 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547348 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547363 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547379 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547395 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547410 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547425 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547440 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547455 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547470 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547484 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547500 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547515 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547531 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547546 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547562 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547580 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547596 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547613 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547628 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547646 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547662 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547690 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547706 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547737 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547753 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547769 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547786 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547806 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547821 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547838 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547854 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547885 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547902 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547919 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547936 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547952 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547967 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547983 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.547999 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548014 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548030 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548046 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548062 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548078 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548095 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548110 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548127 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548147 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548166 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548181 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548198 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548244 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548260 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548276 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548294 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548311 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548350 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548367 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548383 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548402 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548417 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548434 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548453 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548469 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548484 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548501 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548517 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548533 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548550 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548566 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548573 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548582 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548630 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548649 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548668 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548685 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548701 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548718 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548734 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548747 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548752 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548806 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548829 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548853 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548889 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548908 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548980 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.548990 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549004 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549032 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549059 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549079 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549098 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549116 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549134 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549156 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549175 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549180 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549216 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549236 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549255 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549311 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549323 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549335 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549347 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549621 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.549826 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.550041 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.550219 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.550328 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.550399 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.550599 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.550724 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.550772 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.550975 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551095 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551241 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551346 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551402 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551535 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551569 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551729 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551734 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551738 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551804 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.551887 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552061 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552115 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552183 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552295 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552467 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552516 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552546 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552805 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552929 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.553070 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.553106 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.553133 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.552828 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.553428 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.553857 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559172 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559214 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559250 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559434 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559470 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559537 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559580 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559735 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559755 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.559847 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.560012 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.560106 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.560120 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.560225 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.560398 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.560409 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.560426 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.562340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.562405 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.562567 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.562581 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.563229 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.563436 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.563634 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.563735 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.563832 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:08:32.06381311 +0000 UTC m=+20.212854066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564140 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564146 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564200 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564237 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564268 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564334 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564480 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564800 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.564962 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565304 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565423 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565425 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565526 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565574 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565608 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565630 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565718 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565797 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.565804 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.566008 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.566031 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.566175 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.566355 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.566378 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.566657 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.566711 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.566815 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.567096 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.567100 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.567242 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.567318 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.567408 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.567514 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.567700 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.567700 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.568011 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.568292 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.568333 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.568640 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.568655 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.568648 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.568849 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.568860 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573104 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573122 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573297 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573319 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573468 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573469 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573537 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.573570 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.573629 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:32.073612979 +0000 UTC m=+20.222653935 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573664 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573696 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573834 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.573858 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.574036 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.574049 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.574078 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.574188 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.574241 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.574486 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.574666 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.575573 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.575611 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.575939 4726 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.575975 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.576085 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.576233 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.576267 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.576507 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.576516 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.576734 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.576734 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.577049 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.577150 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.577232 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.577573 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.577912 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.578084 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.578461 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.578619 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.578795 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.578980 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.579169 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.579341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.579515 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.579770 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.579970 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.580146 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.580283 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.583585 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.583900 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.584259 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.584542 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.584550 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.584832 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.584949 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.585010 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.585467 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.585960 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.586104 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.586276 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.586502 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.586777 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.587341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.587379 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.586856 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.587758 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.587776 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.587836 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.588108 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.588158 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:32.088144819 +0000 UTC m=+20.237185765 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.588294 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.588417 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.588666 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.588841 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.589124 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.589755 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.593980 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.594097 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.594320 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.594445 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.594643 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.595020 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.595629 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.597341 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.598160 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.599932 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.607432 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.607617 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.610559 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.614129 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.616417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.617170 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.619952 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.626705 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.626728 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.626740 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.626803 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:32.126788244 +0000 UTC m=+20.275829200 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.635109 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.635166 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.635180 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:31 crc kubenswrapper[4726]: E1123 20:08:31.635238 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:32.135219486 +0000 UTC m=+20.284260442 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.635466 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.639505 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.644908 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.647920 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.650986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651148 4726 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651164 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651173 4726 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651182 4726 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651190 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651198 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651206 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651215 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651222 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651231 4726 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651239 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651247 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651256 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651265 4726 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651274 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651282 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651290 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651298 4726 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651306 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651316 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651324 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651333 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651341 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651349 4726 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651357 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651365 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651373 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651381 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651390 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651398 4726 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651406 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651414 4726 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651422 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651431 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651439 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651447 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651456 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651464 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651472 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651480 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651488 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651497 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651504 4726 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651514 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651523 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651531 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651540 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651548 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651556 4726 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651564 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651572 4726 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651580 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651588 4726 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651597 4726 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651605 4726 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651614 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651622 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651630 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651639 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651647 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651655 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651663 4726 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651671 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651680 4726 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651688 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651696 4726 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651705 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651713 4726 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651722 4726 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651730 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651739 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651748 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651756 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651766 4726 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651774 4726 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651787 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651795 4726 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651804 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651812 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651820 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651828 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651835 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651844 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651852 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651860 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651881 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651891 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651899 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651907 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651915 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651923 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651940 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651948 4726 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651956 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651966 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651974 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651982 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.651995 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652003 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652011 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652019 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652028 4726 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652035 4726 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652043 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652051 4726 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652060 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652068 4726 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652076 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652083 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652092 4726 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652100 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652108 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652117 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652125 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652136 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652145 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652155 4726 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652163 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652171 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652179 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652187 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652195 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652203 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652211 4726 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652221 4726 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652228 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652236 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652244 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652252 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652262 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652269 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652277 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652285 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652294 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652301 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652309 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652317 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652325 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652333 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652346 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652355 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652363 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652371 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652378 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652386 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652394 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652402 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652410 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652418 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652426 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652434 4726 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652442 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652451 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652459 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652469 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652478 4726 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652485 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652493 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652501 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652509 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652517 4726 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652527 4726 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652535 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652543 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652550 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652559 4726 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652567 4726 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652575 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652582 4726 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652590 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652598 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652606 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652613 4726 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652622 4726 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652630 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652638 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652645 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652653 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652661 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652669 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652677 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652686 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652695 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652704 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652712 4726 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652720 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.652729 4726 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.653191 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.653292 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.662165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.662678 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.672637 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.679793 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.693094 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.706837 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-bq9lq"] Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.707167 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bq9lq" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.709342 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.709366 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.709901 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.712551 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.728335 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.735063 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.736583 4726 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3" exitCode=255 Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.737119 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3"} Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.743932 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.751716 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.751987 4726 scope.go:117] "RemoveContainer" containerID="9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.753632 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c99d511-e691-4e43-a4cd-fbd23a033b92-hosts-file\") pod \"node-resolver-bq9lq\" (UID: \"7c99d511-e691-4e43-a4cd-fbd23a033b92\") " pod="openshift-dns/node-resolver-bq9lq" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.753702 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfz7n\" (UniqueName: \"kubernetes.io/projected/7c99d511-e691-4e43-a4cd-fbd23a033b92-kube-api-access-vfz7n\") pod \"node-resolver-bq9lq\" (UID: \"7c99d511-e691-4e43-a4cd-fbd23a033b92\") " pod="openshift-dns/node-resolver-bq9lq" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.753739 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.753752 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.753761 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.755331 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.774223 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.783236 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.797286 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.801572 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.815046 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.824661 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.830354 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.833343 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.843520 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.843766 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.854191 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfz7n\" (UniqueName: \"kubernetes.io/projected/7c99d511-e691-4e43-a4cd-fbd23a033b92-kube-api-access-vfz7n\") pod \"node-resolver-bq9lq\" (UID: \"7c99d511-e691-4e43-a4cd-fbd23a033b92\") " pod="openshift-dns/node-resolver-bq9lq" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.854248 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c99d511-e691-4e43-a4cd-fbd23a033b92-hosts-file\") pod \"node-resolver-bq9lq\" (UID: \"7c99d511-e691-4e43-a4cd-fbd23a033b92\") " pod="openshift-dns/node-resolver-bq9lq" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.854319 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7c99d511-e691-4e43-a4cd-fbd23a033b92-hosts-file\") pod \"node-resolver-bq9lq\" (UID: \"7c99d511-e691-4e43-a4cd-fbd23a033b92\") " pod="openshift-dns/node-resolver-bq9lq" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.860922 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: W1123 20:08:31.863045 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-e70e1b00154728d7df20a5be989684057615972f825ca8485b75f45f6df4fc7e WatchSource:0}: Error finding container e70e1b00154728d7df20a5be989684057615972f825ca8485b75f45f6df4fc7e: Status 404 returned error can't find the container with id e70e1b00154728d7df20a5be989684057615972f825ca8485b75f45f6df4fc7e Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.872316 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfz7n\" (UniqueName: \"kubernetes.io/projected/7c99d511-e691-4e43-a4cd-fbd23a033b92-kube-api-access-vfz7n\") pod \"node-resolver-bq9lq\" (UID: \"7c99d511-e691-4e43-a4cd-fbd23a033b92\") " pod="openshift-dns/node-resolver-bq9lq" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.877338 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: W1123 20:08:31.881621 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-323af1289ca8476e36f2969f13fda31a3bf521c91e30277c86b4ed110d21931a WatchSource:0}: Error finding container 323af1289ca8476e36f2969f13fda31a3bf521c91e30277c86b4ed110d21931a: Status 404 returned error can't find the container with id 323af1289ca8476e36f2969f13fda31a3bf521c91e30277c86b4ed110d21931a Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.888179 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.900276 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.918597 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.935296 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:31 crc kubenswrapper[4726]: I1123 20:08:31.947086 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.018486 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bq9lq" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.157409 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.157476 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.157501 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.157520 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.157540 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157636 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157651 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157661 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157696 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:33.157684802 +0000 UTC m=+21.306725758 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157741 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:08:33.157735524 +0000 UTC m=+21.306776470 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157768 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157791 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:33.157786015 +0000 UTC m=+21.306826971 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157838 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157847 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157854 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157892 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:33.157886418 +0000 UTC m=+21.306927374 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157930 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:32 crc kubenswrapper[4726]: E1123 20:08:32.157948 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:33.157943459 +0000 UTC m=+21.306984415 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.593180 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.594047 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.594815 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.595397 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.595982 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.596429 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.597010 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.597534 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.598185 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.598844 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.600892 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.601690 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.602540 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.603048 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.603861 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.604352 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.605220 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.605625 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.608290 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.608440 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.609703 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.610211 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.610804 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.611259 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.612723 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.613196 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.613792 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.614999 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.615449 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.616462 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.616977 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.620365 4726 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.620479 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.622428 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.623054 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.623503 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.625231 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.629497 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.630082 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.630594 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.631077 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.631705 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.632516 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.633106 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.634162 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.635149 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.635600 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.636467 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.636992 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.638250 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.638776 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.639425 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.640431 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.641117 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.642168 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.642633 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.642841 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.659041 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.681215 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.694469 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.711225 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.720654 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.738281 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.740345 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bq9lq" event={"ID":"7c99d511-e691-4e43-a4cd-fbd23a033b92","Type":"ContainerStarted","Data":"ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.740410 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bq9lq" event={"ID":"7c99d511-e691-4e43-a4cd-fbd23a033b92","Type":"ContainerStarted","Data":"8b34c524b91af59756982fe649f2e80394fc7b3537dd78405d64cf20166e8779"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.741912 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.741979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.741993 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"323af1289ca8476e36f2969f13fda31a3bf521c91e30277c86b4ed110d21931a"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.743270 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.743296 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e70e1b00154728d7df20a5be989684057615972f825ca8485b75f45f6df4fc7e"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.743957 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a604903a0638d75bedd8d4e0b77f3bc72e20b6a231f8e740aac0a44f47ba8597"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.747260 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.752091 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c"} Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.757245 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.768731 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.799933 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.836300 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.849099 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.861085 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.873632 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.886038 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.899712 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.920482 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.933577 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.946922 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.957421 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.975489 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:32 crc kubenswrapper[4726]: I1123 20:08:32.988073 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.015067 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.029746 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.049222 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.141565 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-jttr8"] Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.141861 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.142309 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-c58qk"] Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.142833 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.146264 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.146375 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.146448 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.146915 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.146958 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.157250 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.162322 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.162505 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.162511 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.162614 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.162711 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-td7h8"] Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.164649 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.166487 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.166568 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.166621 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.166648 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.166676 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167191 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167229 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167247 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167355 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:08:35.167334993 +0000 UTC m=+23.316375949 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167375 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167409 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167444 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:35.167423445 +0000 UTC m=+23.316464401 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167470 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:35.167458776 +0000 UTC m=+23.316499732 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167509 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:35.167479257 +0000 UTC m=+23.316520213 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167580 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167606 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167625 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.167681 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:35.167664912 +0000 UTC m=+23.316705868 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.182389 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.182569 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-mpwz9"] Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.182973 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.183316 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.186596 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.186688 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.186921 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.187368 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.192150 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.192424 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.201105 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.243742 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268075 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-systemd\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268123 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-etc-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268142 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268209 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-systemd-units\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268279 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-netns\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268307 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rlzx\" (UniqueName: \"kubernetes.io/projected/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-kube-api-access-2rlzx\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268335 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e3ac186-9f76-4774-8e04-fb00add1eb72-mcd-auth-proxy-config\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268355 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-os-release\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268387 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-etc-kubernetes\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268479 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-var-lib-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268551 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268576 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-config\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268594 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-multus-certs\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268624 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-system-cni-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268642 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-k8s-cni-cncf-io\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268659 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268702 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-bin\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268719 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e3ac186-9f76-4774-8e04-fb00add1eb72-proxy-tls\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268738 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-cni-bin\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268778 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf84x\" (UniqueName: \"kubernetes.io/projected/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-kube-api-access-zf84x\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268803 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-cni-multus\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268826 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovn-node-metrics-cert\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268855 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcgdp\" (UniqueName: \"kubernetes.io/projected/2e3ac186-9f76-4774-8e04-fb00add1eb72-kube-api-access-tcgdp\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268888 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-cnibin\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268908 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-cni-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268927 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-system-cni-dir\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268948 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.268990 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-node-log\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269029 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-netd\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269046 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-env-overrides\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269063 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2e3ac186-9f76-4774-8e04-fb00add1eb72-rootfs\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269083 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cnibin\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269105 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8c5753d1-1307-4d28-b08f-3588c14ec1bd-cni-binary-copy\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269121 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-conf-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269139 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-474m4\" (UniqueName: \"kubernetes.io/projected/8c5753d1-1307-4d28-b08f-3588c14ec1bd-kube-api-access-474m4\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269155 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-slash\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269172 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-netns\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269207 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-kubelet\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269267 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cni-binary-copy\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269318 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-script-lib\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269343 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-os-release\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269362 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-socket-dir-parent\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269381 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-hostroot\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269411 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-kubelet\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269434 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-ovn\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269453 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-ovn-kubernetes\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269491 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-log-socket\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.269513 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-daemon-config\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.270127 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.296054 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.327589 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.351296 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370408 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-kubelet\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370452 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-ovn\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370474 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-ovn-kubernetes\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370499 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-log-socket\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370518 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-daemon-config\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370536 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-systemd\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370541 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-kubelet\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370590 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-etc-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370594 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-log-socket\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370554 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-etc-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370615 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-ovn-kubernetes\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370633 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-systemd\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-ovn\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370851 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-systemd-units\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370901 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-systemd-units\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370906 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370945 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-netns\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370970 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rlzx\" (UniqueName: \"kubernetes.io/projected/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-kube-api-access-2rlzx\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370992 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e3ac186-9f76-4774-8e04-fb00add1eb72-mcd-auth-proxy-config\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.370995 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-netns\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371009 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-os-release\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371026 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-var-lib-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371042 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371058 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-config\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371074 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-multus-certs\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371089 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-etc-kubernetes\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371106 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-system-cni-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371125 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-k8s-cni-cncf-io\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371141 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371140 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-var-lib-openvswitch\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371163 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-bin\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371184 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e3ac186-9f76-4774-8e04-fb00add1eb72-proxy-tls\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371207 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-etc-kubernetes\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371229 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371238 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-multus-certs\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371266 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-run-k8s-cni-cncf-io\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371305 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-cni-bin\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371368 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-bin\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371375 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf84x\" (UniqueName: \"kubernetes.io/projected/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-kube-api-access-zf84x\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371405 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-daemon-config\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371407 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-cni-multus\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371454 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovn-node-metrics-cert\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371473 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcgdp\" (UniqueName: \"kubernetes.io/projected/2e3ac186-9f76-4774-8e04-fb00add1eb72-kube-api-access-tcgdp\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371490 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-cnibin\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371509 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-cni-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371526 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-system-cni-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371527 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-system-cni-dir\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371551 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-system-cni-dir\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371590 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-cnibin\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371606 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-cni-multus\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371637 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-cni-bin\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371667 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-cni-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371671 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e3ac186-9f76-4774-8e04-fb00add1eb72-mcd-auth-proxy-config\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371697 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371800 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-node-log\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-config\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371849 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-node-log\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371921 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-netd\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371951 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-env-overrides\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371965 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-netd\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.371999 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2e3ac186-9f76-4774-8e04-fb00add1eb72-rootfs\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372085 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2e3ac186-9f76-4774-8e04-fb00add1eb72-rootfs\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372126 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8c5753d1-1307-4d28-b08f-3588c14ec1bd-cni-binary-copy\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372145 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-conf-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372161 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-474m4\" (UniqueName: \"kubernetes.io/projected/8c5753d1-1307-4d28-b08f-3588c14ec1bd-kube-api-access-474m4\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372206 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cnibin\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372225 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-slash\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372259 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-netns\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372275 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-kubelet\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372290 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cni-binary-copy\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372307 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-script-lib\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372339 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-os-release\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372355 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-socket-dir-parent\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372359 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-env-overrides\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372369 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-hostroot\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372384 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-conf-dir\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372397 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-host-var-lib-kubelet\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372436 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-netns\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372445 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-hostroot\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372458 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cnibin\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372490 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-multus-socket-dir-parent\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372510 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-slash\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372529 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8c5753d1-1307-4d28-b08f-3588c14ec1bd-os-release\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372558 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8c5753d1-1307-4d28-b08f-3588c14ec1bd-cni-binary-copy\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.372597 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-os-release\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.373051 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cni-binary-copy\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.374934 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.376540 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovn-node-metrics-cert\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.381365 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-script-lib\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.384491 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e3ac186-9f76-4774-8e04-fb00add1eb72-proxy-tls\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.389982 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.395431 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rlzx\" (UniqueName: \"kubernetes.io/projected/7a041870-99e1-48bc-99a6-f1c3c01fa5ce-kube-api-access-2rlzx\") pod \"multus-additional-cni-plugins-mpwz9\" (UID: \"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\") " pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.405476 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-474m4\" (UniqueName: \"kubernetes.io/projected/8c5753d1-1307-4d28-b08f-3588c14ec1bd-kube-api-access-474m4\") pod \"multus-jttr8\" (UID: \"8c5753d1-1307-4d28-b08f-3588c14ec1bd\") " pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.411452 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf84x\" (UniqueName: \"kubernetes.io/projected/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-kube-api-access-zf84x\") pod \"ovnkube-node-td7h8\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.411477 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcgdp\" (UniqueName: \"kubernetes.io/projected/2e3ac186-9f76-4774-8e04-fb00add1eb72-kube-api-access-tcgdp\") pod \"machine-config-daemon-c58qk\" (UID: \"2e3ac186-9f76-4774-8e04-fb00add1eb72\") " pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.419379 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.432999 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.449459 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.453518 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jttr8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.460085 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:08:33 crc kubenswrapper[4726]: W1123 20:08:33.470479 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c5753d1_1307_4d28_b08f_3588c14ec1bd.slice/crio-2d33961852e195f9f9eb3d37e685c5597947461fd679da0ddab8b346d2e56587 WatchSource:0}: Error finding container 2d33961852e195f9f9eb3d37e685c5597947461fd679da0ddab8b346d2e56587: Status 404 returned error can't find the container with id 2d33961852e195f9f9eb3d37e685c5597947461fd679da0ddab8b346d2e56587 Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.476449 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.476498 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.494052 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.499395 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: W1123 20:08:33.499753 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb0ef36c_6b04_4a1f_bb8d_8f0895fd33be.slice/crio-504f88d9cf2135f28529f93c72c6e5ae0ca125dec7ee9de049b05ed7d5c29c7a WatchSource:0}: Error finding container 504f88d9cf2135f28529f93c72c6e5ae0ca125dec7ee9de049b05ed7d5c29c7a: Status 404 returned error can't find the container with id 504f88d9cf2135f28529f93c72c6e5ae0ca125dec7ee9de049b05ed7d5c29c7a Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.514299 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.532372 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.556344 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.574475 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.590067 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.590226 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.590491 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.590613 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.590654 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:33 crc kubenswrapper[4726]: E1123 20:08:33.590709 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.593980 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.627622 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.640912 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.661859 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.674962 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.691750 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.707896 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.737207 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.756516 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jttr8" event={"ID":"8c5753d1-1307-4d28-b08f-3588c14ec1bd","Type":"ContainerStarted","Data":"0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec"} Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.756568 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jttr8" event={"ID":"8c5753d1-1307-4d28-b08f-3588c14ec1bd","Type":"ContainerStarted","Data":"2d33961852e195f9f9eb3d37e685c5597947461fd679da0ddab8b346d2e56587"} Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.758331 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerStarted","Data":"82da7a1d64fae8b47246c67a6fcef2d091109969535878c991d0e5c66e475c34"} Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.760556 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e" exitCode=0 Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.760627 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.760656 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"504f88d9cf2135f28529f93c72c6e5ae0ca125dec7ee9de049b05ed7d5c29c7a"} Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.764421 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60"} Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.764458 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"cd234612f554beae75a2c056348ed6f32c2ab60249a79853bdece916abe1f0e4"} Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.764771 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.772264 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.788846 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.807915 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.828576 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.858710 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.872905 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.896081 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.919640 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.937760 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:33 crc kubenswrapper[4726]: I1123 20:08:33.953542 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:33Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.014678 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.018328 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.025590 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.040368 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.043090 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.059569 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.072970 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.093016 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.114743 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.147700 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.172592 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.201968 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.220147 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.265708 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.283429 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.305818 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.323427 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.347164 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.384613 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.425831 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.768822 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.770032 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.771311 4726 generic.go:334] "Generic (PLEG): container finished" podID="7a041870-99e1-48bc-99a6-f1c3c01fa5ce" containerID="90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec" exitCode=0 Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.771381 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerDied","Data":"90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.775863 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.775951 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.775968 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.775982 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.775997 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.776013 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.785126 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.793238 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-p67p9"] Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.793719 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.797027 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.797332 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.797408 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.799363 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.810380 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.827386 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.842442 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.856560 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.885909 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.891684 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/90959adf-c835-40fa-b0f1-26e80426a5c8-serviceca\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.891738 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rg5t\" (UniqueName: \"kubernetes.io/projected/90959adf-c835-40fa-b0f1-26e80426a5c8-kube-api-access-8rg5t\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.891809 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90959adf-c835-40fa-b0f1-26e80426a5c8-host\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.906710 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.925201 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.941107 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.956582 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.970029 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.986240 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:34Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.993680 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/90959adf-c835-40fa-b0f1-26e80426a5c8-serviceca\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.993728 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rg5t\" (UniqueName: \"kubernetes.io/projected/90959adf-c835-40fa-b0f1-26e80426a5c8-kube-api-access-8rg5t\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.993777 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90959adf-c835-40fa-b0f1-26e80426a5c8-host\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.993853 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90959adf-c835-40fa-b0f1-26e80426a5c8-host\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:34 crc kubenswrapper[4726]: I1123 20:08:34.994742 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/90959adf-c835-40fa-b0f1-26e80426a5c8-serviceca\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.032373 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rg5t\" (UniqueName: \"kubernetes.io/projected/90959adf-c835-40fa-b0f1-26e80426a5c8-kube-api-access-8rg5t\") pod \"node-ca-p67p9\" (UID: \"90959adf-c835-40fa-b0f1-26e80426a5c8\") " pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.045641 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.087887 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.126243 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.166081 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.194905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.195032 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.195064 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195115 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:08:39.195066688 +0000 UTC m=+27.344107654 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195163 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195186 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195205 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195211 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:39.195199671 +0000 UTC m=+27.344240627 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.195208 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195217 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195370 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:39.195351426 +0000 UTC m=+27.344392392 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195271 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195406 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195420 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195458 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:39.195449268 +0000 UTC m=+27.344490234 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.195478 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195543 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.195613 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:39.195596882 +0000 UTC m=+27.344637848 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.206917 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.246224 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.294416 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.327348 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.352806 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-p67p9" Nov 23 20:08:35 crc kubenswrapper[4726]: W1123 20:08:35.369003 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90959adf_c835_40fa_b0f1_26e80426a5c8.slice/crio-a24b291e042594deb9a6a5413898c77d9931504c54617ecda540a3b45fdbcbb3 WatchSource:0}: Error finding container a24b291e042594deb9a6a5413898c77d9931504c54617ecda540a3b45fdbcbb3: Status 404 returned error can't find the container with id a24b291e042594deb9a6a5413898c77d9931504c54617ecda540a3b45fdbcbb3 Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.376719 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.410842 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.449301 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.489083 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.531984 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.564553 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.592052 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.592176 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.592477 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.592533 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.592572 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:35 crc kubenswrapper[4726]: E1123 20:08:35.592619 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.607223 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.653744 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.724106 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.779928 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-p67p9" event={"ID":"90959adf-c835-40fa-b0f1-26e80426a5c8","Type":"ContainerStarted","Data":"eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c"} Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.779968 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-p67p9" event={"ID":"90959adf-c835-40fa-b0f1-26e80426a5c8","Type":"ContainerStarted","Data":"a24b291e042594deb9a6a5413898c77d9931504c54617ecda540a3b45fdbcbb3"} Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.781814 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerStarted","Data":"f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5"} Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.793497 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.809894 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.822237 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.846339 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.884598 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.925324 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:35 crc kubenswrapper[4726]: I1123 20:08:35.964177 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:35Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.011734 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.049388 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.085551 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.124629 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.163834 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.205778 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.247201 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.286629 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.323599 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.379957 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.405756 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.447490 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.506092 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.534969 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.571994 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.608291 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.646600 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.686583 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.726413 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.770355 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.786775 4726 generic.go:334] "Generic (PLEG): container finished" podID="7a041870-99e1-48bc-99a6-f1c3c01fa5ce" containerID="f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5" exitCode=0 Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.786884 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerDied","Data":"f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5"} Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.792285 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.825255 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.849687 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.884746 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.925956 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:36 crc kubenswrapper[4726]: I1123 20:08:36.966259 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:36Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.008097 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.044334 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.093602 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.125186 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.169486 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.210507 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.249458 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.291482 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.327388 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.377886 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.410299 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.446532 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.490119 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.588862 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.588965 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.588996 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.588967 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.589099 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.589153 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.798726 4726 generic.go:334] "Generic (PLEG): container finished" podID="7a041870-99e1-48bc-99a6-f1c3c01fa5ce" containerID="d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448" exitCode=0 Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.798802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerDied","Data":"d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448"} Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.826029 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.844028 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.847035 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.849223 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.849283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.849303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.849489 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.858993 4726 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.859463 4726 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.865489 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.868748 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.868843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.868862 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.868960 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.868983 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:37Z","lastTransitionTime":"2025-11-23T20:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.884729 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.889469 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.894338 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.894385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.894399 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.894421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.894449 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:37Z","lastTransitionTime":"2025-11-23T20:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.908082 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.909206 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.914895 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.914946 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.914972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.915005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.915028 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:37Z","lastTransitionTime":"2025-11-23T20:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.925351 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.928466 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.934399 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.934438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.934447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.934464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.934474 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:37Z","lastTransitionTime":"2025-11-23T20:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.941739 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.947128 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.950310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.950341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.950352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.950366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.950377 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:37Z","lastTransitionTime":"2025-11-23T20:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.962715 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.963657 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:37 crc kubenswrapper[4726]: E1123 20:08:37.963795 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.966403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.966450 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.966462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.966482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.966495 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:37Z","lastTransitionTime":"2025-11-23T20:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:37 crc kubenswrapper[4726]: I1123 20:08:37.976244 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.001974 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:37Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.017334 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.027174 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.044090 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.069418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.069454 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.069466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.069483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.069494 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.087776 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.122740 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.171994 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.172038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.172049 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.172066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.172077 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.274457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.275423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.275501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.275576 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.275635 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.379214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.379461 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.379479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.379505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.379524 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.483372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.483427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.483444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.483472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.483490 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.586295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.586357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.586376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.586401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.586420 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.689543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.689789 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.689944 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.690049 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.690139 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.793587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.793685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.794203 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.794291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.794576 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.806948 4726 generic.go:334] "Generic (PLEG): container finished" podID="7a041870-99e1-48bc-99a6-f1c3c01fa5ce" containerID="4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d" exitCode=0 Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.806990 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerDied","Data":"4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.842631 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.861650 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.876729 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.892956 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.899024 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.899065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.899083 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.899108 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.899125 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:38Z","lastTransitionTime":"2025-11-23T20:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.907686 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.928994 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.945264 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.961863 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.976746 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:38 crc kubenswrapper[4726]: I1123 20:08:38.996500 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:38Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.001474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.001524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.001562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.001594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.001606 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.011270 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.031130 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.054049 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.067148 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.087097 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.104252 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.104279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.104289 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.104305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.104316 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.206443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.206476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.206488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.206503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.206513 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.242531 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.242657 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.242721 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.242758 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.242804 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243010 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243076 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243023 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243225 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243240 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243256 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243266 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243285 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243084 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:47.243067591 +0000 UTC m=+35.392108537 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243356 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:08:47.243328248 +0000 UTC m=+35.392369244 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243381 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:47.243369089 +0000 UTC m=+35.392410075 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243407 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:47.24339698 +0000 UTC m=+35.392437976 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.243431 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:47.243418291 +0000 UTC m=+35.392459277 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.309981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.310024 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.310100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.310120 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.310132 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.413360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.413394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.413406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.413422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.413435 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.515974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.516005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.516015 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.516028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.516038 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.588517 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.588566 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.588862 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.588928 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.588580 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:39 crc kubenswrapper[4726]: E1123 20:08:39.588980 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.619048 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.619088 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.619100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.619114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.619125 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.721324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.721362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.721373 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.721389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.721399 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.813404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerStarted","Data":"af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.819399 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.819732 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.819769 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.824319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.824363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.824419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.824443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.824461 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.836315 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.851165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.853365 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.853902 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.870056 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.882250 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.898750 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.915921 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.927347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.927510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.927600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.927709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.927809 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:39Z","lastTransitionTime":"2025-11-23T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.930084 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.949926 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.969926 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:39 crc kubenswrapper[4726]: I1123 20:08:39.991115 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:39Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.004502 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.030391 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.031193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.031235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.031251 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.031274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.031292 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.058857 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.079003 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.098574 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.112056 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.125608 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.134147 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.134201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.134222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.134372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.134392 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.146070 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.168738 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.187683 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.205393 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.225733 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.237397 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.237457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.237482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.237513 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.237536 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.249107 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.266774 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.300526 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.316382 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.335347 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.339625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.339703 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.339743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.339776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.339798 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.356284 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.371327 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.403165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.442109 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.442174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.442198 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.442225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.442246 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.545036 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.545090 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.545109 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.545131 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.545148 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.648571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.648636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.648659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.648691 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.648712 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.751193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.751250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.751267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.751290 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.751307 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.824893 4726 generic.go:334] "Generic (PLEG): container finished" podID="7a041870-99e1-48bc-99a6-f1c3c01fa5ce" containerID="af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4" exitCode=0 Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.824938 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerDied","Data":"af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.825024 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.836634 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.853693 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.854688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.854734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.854745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.854763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.854776 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.866299 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.878150 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.898952 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.918776 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.934698 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.947170 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.958832 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.963939 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.963971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.963981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.963996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.964006 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:40Z","lastTransitionTime":"2025-11-23T20:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.975533 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:40 crc kubenswrapper[4726]: I1123 20:08:40.989020 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.001466 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:40Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.013588 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.022725 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.042303 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.066051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.066079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.066088 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.066102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.066112 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.168553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.168583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.168591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.168607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.168616 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.270329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.270386 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.270412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.270428 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.270439 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.372213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.372241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.372250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.372262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.372271 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.474575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.474607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.474617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.474629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.474637 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.576855 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.576905 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.576922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.576937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.576948 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.588672 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.588672 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:41 crc kubenswrapper[4726]: E1123 20:08:41.588831 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:41 crc kubenswrapper[4726]: E1123 20:08:41.588770 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.588672 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:41 crc kubenswrapper[4726]: E1123 20:08:41.588908 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.679319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.679350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.679358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.679370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.679380 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.781498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.781536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.781546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.781563 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.781572 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.831760 4726 generic.go:334] "Generic (PLEG): container finished" podID="7a041870-99e1-48bc-99a6-f1c3c01fa5ce" containerID="668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00" exitCode=0 Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.831833 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerDied","Data":"668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.833650 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/0.log" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.838311 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c" exitCode=1 Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.838353 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.839122 4726 scope.go:117] "RemoveContainer" containerID="8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.843059 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.874056 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.891364 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.891421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.891446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.891467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.891485 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.891499 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.907437 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.918562 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.933488 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.946388 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.963684 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.974255 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.993562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.993585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.993593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.993607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.993629 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:41Z","lastTransitionTime":"2025-11-23T20:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:41 crc kubenswrapper[4726]: I1123 20:08:41.999783 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:41Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.011551 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.027191 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.039990 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.050819 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.061703 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.073667 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.084770 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.094715 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.096066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.096097 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.096146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.096164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.096397 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.105056 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.116367 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.126999 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.142765 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.153770 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.173658 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.185122 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.197748 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.199250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.199294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.199304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.199319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.199328 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.210245 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.222662 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.232764 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.254945 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"message\\\":\\\"ect:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:08:41.723223 5870 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI1123 20:08:41.723230 5870 services_controller.go:452] Built service openshift-network-diagnostics/network-check-target per-node LB for network=default: []services.LB{}\\\\nI1123 20:08:41.723205 5870 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc009839ad0] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI1123 20:08:41.723241 5870 services_controller.go:453] Built service openshift-network-diagnostics/network-check-target template LB for network=default: []services.LB{}\\\\nI1123 20:08:41.723246 5870 pods.go:252] [openshift-network-console/networking-conso\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.301925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.302200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.302444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.302646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.302844 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.404718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.404763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.404771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.404787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.404796 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.506908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.506953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.506963 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.506976 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.506986 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.607520 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.609568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.609604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.609617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.609635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.609648 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.624805 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.636106 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.660079 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.672814 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.692227 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.710162 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.711242 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.711272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.711281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.711296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.711306 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.721494 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.744174 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.760188 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.776094 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.786930 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.801430 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.813401 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.814456 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.814509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.814526 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.814548 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.814565 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.832175 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"message\\\":\\\"ect:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:08:41.723223 5870 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI1123 20:08:41.723230 5870 services_controller.go:452] Built service openshift-network-diagnostics/network-check-target per-node LB for network=default: []services.LB{}\\\\nI1123 20:08:41.723205 5870 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc009839ad0] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI1123 20:08:41.723241 5870 services_controller.go:453] Built service openshift-network-diagnostics/network-check-target template LB for network=default: []services.LB{}\\\\nI1123 20:08:41.723246 5870 pods.go:252] [openshift-network-console/networking-conso\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.845406 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" event={"ID":"7a041870-99e1-48bc-99a6-f1c3c01fa5ce","Type":"ContainerStarted","Data":"78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.847922 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/1.log" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.849065 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/0.log" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.852630 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737" exitCode=1 Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.852677 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.852732 4726 scope.go:117] "RemoveContainer" containerID="8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.854737 4726 scope.go:117] "RemoveContainer" containerID="86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737" Nov 23 20:08:42 crc kubenswrapper[4726]: E1123 20:08:42.855121 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.866491 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.882580 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.898358 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.912277 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.916638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.916659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.916690 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.916704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.916713 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:42Z","lastTransitionTime":"2025-11-23T20:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.932938 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.944645 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.957140 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.972636 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.982861 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:42 crc kubenswrapper[4726]: I1123 20:08:42.999172 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:42Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.008329 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.020322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.020367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.020384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.020407 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.020424 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.034632 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"message\\\":\\\"ect:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:08:41.723223 5870 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI1123 20:08:41.723230 5870 services_controller.go:452] Built service openshift-network-diagnostics/network-check-target per-node LB for network=default: []services.LB{}\\\\nI1123 20:08:41.723205 5870 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc009839ad0] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI1123 20:08:41.723241 5870 services_controller.go:453] Built service openshift-network-diagnostics/network-check-target template LB for network=default: []services.LB{}\\\\nI1123 20:08:41.723246 5870 pods.go:252] [openshift-network-console/networking-conso\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.046360 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.059716 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.073652 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.090814 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.106316 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.122813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.122892 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.122912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.122949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.122969 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.125535 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.136672 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.168805 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.191424 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.204341 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.227287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.227592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.227716 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.227835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.228024 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.246845 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.285406 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.324475 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.329986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.330133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.330230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.330331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.330427 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.362816 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.426402 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c3fbea9751c6fcf6fc0f5d8afd91a238983c06547801f54d48cb550e6153a8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"message\\\":\\\"ect:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:08:41.723223 5870 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI1123 20:08:41.723230 5870 services_controller.go:452] Built service openshift-network-diagnostics/network-check-target per-node LB for network=default: []services.LB{}\\\\nI1123 20:08:41.723205 5870 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc009839ad0] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI1123 20:08:41.723241 5870 services_controller.go:453] Built service openshift-network-diagnostics/network-check-target template LB for network=default: []services.LB{}\\\\nI1123 20:08:41.723246 5870 pods.go:252] [openshift-network-console/networking-conso\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.433080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.433145 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.433166 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.433199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.433220 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.459334 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.496900 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.525059 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.545217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.545276 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.545291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.545312 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.545323 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.588999 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.589083 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.589090 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:43 crc kubenswrapper[4726]: E1123 20:08:43.589428 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:43 crc kubenswrapper[4726]: E1123 20:08:43.589611 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:43 crc kubenswrapper[4726]: E1123 20:08:43.589638 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.649164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.649219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.649230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.649246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.649260 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.752557 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.752684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.752695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.752710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.752719 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.856543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.856583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.856592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.856607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.856617 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.858053 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/1.log" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.863142 4726 scope.go:117] "RemoveContainer" containerID="86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737" Nov 23 20:08:43 crc kubenswrapper[4726]: E1123 20:08:43.863454 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.882577 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.909764 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.924944 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.935218 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.959652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.959742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.959763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.959797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.959820 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:43Z","lastTransitionTime":"2025-11-23T20:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.966808 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.982778 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:43 crc kubenswrapper[4726]: I1123 20:08:43.998549 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:43Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.011230 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:44Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.028641 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:44Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.048839 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:44Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.062008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.062068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.062084 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.062108 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.062121 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.067890 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:44Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.085166 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:44Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.104497 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:44Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.118143 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:44Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.148364 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:44Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.165752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.165837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.165857 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.165923 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.165943 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.270017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.270106 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.270128 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.270150 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.270195 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.373062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.373121 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.373138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.373161 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.373180 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.382951 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.476268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.476370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.476394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.476421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.476441 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.580209 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.580285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.580331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.580368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.580388 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.684206 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.684286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.684308 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.684341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.684365 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.787575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.787611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.787624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.787643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.787656 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.866445 4726 scope.go:117] "RemoveContainer" containerID="86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737" Nov 23 20:08:44 crc kubenswrapper[4726]: E1123 20:08:44.866708 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.890805 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.890927 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.890947 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.891345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.891404 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.994317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.994401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.994418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.994444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:44 crc kubenswrapper[4726]: I1123 20:08:44.994460 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:44Z","lastTransitionTime":"2025-11-23T20:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.097712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.097785 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.097804 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.097833 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.097853 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.201102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.201176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.201194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.201219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.201237 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.303854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.303948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.303963 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.303979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.303990 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.406285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.406345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.406362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.406383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.406398 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.509539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.509584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.509600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.509620 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.509812 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.588862 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.588995 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:45 crc kubenswrapper[4726]: E1123 20:08:45.589109 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:45 crc kubenswrapper[4726]: E1123 20:08:45.589272 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.589508 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:45 crc kubenswrapper[4726]: E1123 20:08:45.589799 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.609451 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k"] Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.610758 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.611742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.611968 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.612155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.612283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.612418 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.616157 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.616174 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.632369 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.651008 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.651444 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw82p\" (UniqueName: \"kubernetes.io/projected/55581d55-59b1-4e06-ac8e-9d7f46b0820b-kube-api-access-mw82p\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.651482 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/55581d55-59b1-4e06-ac8e-9d7f46b0820b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.651529 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/55581d55-59b1-4e06-ac8e-9d7f46b0820b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.651562 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/55581d55-59b1-4e06-ac8e-9d7f46b0820b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.668087 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.684674 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.703574 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.715132 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.715229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.715253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.715282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.715308 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.719950 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.746453 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.752656 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/55581d55-59b1-4e06-ac8e-9d7f46b0820b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.752704 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/55581d55-59b1-4e06-ac8e-9d7f46b0820b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.752739 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw82p\" (UniqueName: \"kubernetes.io/projected/55581d55-59b1-4e06-ac8e-9d7f46b0820b-kube-api-access-mw82p\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.752767 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/55581d55-59b1-4e06-ac8e-9d7f46b0820b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.753319 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/55581d55-59b1-4e06-ac8e-9d7f46b0820b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.755054 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/55581d55-59b1-4e06-ac8e-9d7f46b0820b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.761932 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/55581d55-59b1-4e06-ac8e-9d7f46b0820b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.771512 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.774061 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw82p\" (UniqueName: \"kubernetes.io/projected/55581d55-59b1-4e06-ac8e-9d7f46b0820b-kube-api-access-mw82p\") pod \"ovnkube-control-plane-749d76644c-h665k\" (UID: \"55581d55-59b1-4e06-ac8e-9d7f46b0820b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.783104 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.801784 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.812977 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.818068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.818116 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.818130 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.818149 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.818165 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.827794 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.838337 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.850612 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.862160 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.889470 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:45Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.920206 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.920233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.920246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.920263 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.920274 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:45Z","lastTransitionTime":"2025-11-23T20:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:45 crc kubenswrapper[4726]: I1123 20:08:45.933499 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" Nov 23 20:08:45 crc kubenswrapper[4726]: W1123 20:08:45.950234 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55581d55_59b1_4e06_ac8e_9d7f46b0820b.slice/crio-7e3bee7aca8c1275397a03f84727b5dd5e28a5a6f3da714f7057949cd4bb5449 WatchSource:0}: Error finding container 7e3bee7aca8c1275397a03f84727b5dd5e28a5a6f3da714f7057949cd4bb5449: Status 404 returned error can't find the container with id 7e3bee7aca8c1275397a03f84727b5dd5e28a5a6f3da714f7057949cd4bb5449 Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.022694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.022746 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.022764 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.022786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.022803 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.126697 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.126753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.126783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.126812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.126833 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.229925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.229951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.229961 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.229974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.229983 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.332526 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.332582 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.332595 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.332612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.332643 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.435294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.435331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.435340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.435355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.435364 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.537363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.537401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.537410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.537424 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.537433 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.640286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.640328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.640337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.640351 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.640361 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.721974 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hjmwb"] Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.722592 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:46 crc kubenswrapper[4726]: E1123 20:08:46.722702 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.742831 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.744322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.744386 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.744411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.744442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.744467 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.755791 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.763479 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flz4s\" (UniqueName: \"kubernetes.io/projected/ada6b953-f533-4b33-b07e-7e80604fe4a1-kube-api-access-flz4s\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.763586 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.769410 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.798148 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.811110 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.831364 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.844214 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.847534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.847653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.847679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.847704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.847757 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.863147 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.864466 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flz4s\" (UniqueName: \"kubernetes.io/projected/ada6b953-f533-4b33-b07e-7e80604fe4a1-kube-api-access-flz4s\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.864591 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:46 crc kubenswrapper[4726]: E1123 20:08:46.864766 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:46 crc kubenswrapper[4726]: E1123 20:08:46.864841 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs podName:ada6b953-f533-4b33-b07e-7e80604fe4a1 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:47.364820539 +0000 UTC m=+35.513861525 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs") pod "network-metrics-daemon-hjmwb" (UID: "ada6b953-f533-4b33-b07e-7e80604fe4a1") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.873384 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" event={"ID":"55581d55-59b1-4e06-ac8e-9d7f46b0820b","Type":"ContainerStarted","Data":"b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.873442 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" event={"ID":"55581d55-59b1-4e06-ac8e-9d7f46b0820b","Type":"ContainerStarted","Data":"6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.873458 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" event={"ID":"55581d55-59b1-4e06-ac8e-9d7f46b0820b","Type":"ContainerStarted","Data":"7e3bee7aca8c1275397a03f84727b5dd5e28a5a6f3da714f7057949cd4bb5449"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.883160 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.891276 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flz4s\" (UniqueName: \"kubernetes.io/projected/ada6b953-f533-4b33-b07e-7e80604fe4a1-kube-api-access-flz4s\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.897792 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.914168 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.926660 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.947738 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.950800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.950853 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.950912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.950944 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.950967 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:46Z","lastTransitionTime":"2025-11-23T20:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.960992 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:46 crc kubenswrapper[4726]: I1123 20:08:46.988338 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.000704 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:46Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.020315 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.034945 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.049371 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.054606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.054694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.054721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.054756 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.054782 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.069217 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.082098 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.106117 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.121859 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.140706 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.153272 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.157849 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.157965 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.157992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.158024 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.158046 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.170753 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.187027 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.204014 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.219437 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.232278 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.259798 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.262433 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.262491 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.262508 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.262530 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.262545 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.269791 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.269891 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.269966 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.269969 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:09:03.269936594 +0000 UTC m=+51.418977590 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270058 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:03.270040567 +0000 UTC m=+51.419081563 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.270099 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.270170 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270314 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270331 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270343 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.270340 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270377 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:03.270366956 +0000 UTC m=+51.419407922 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270436 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270449 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270457 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270461 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270482 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:03.270474229 +0000 UTC m=+51.419515195 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.270516 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:03.270501059 +0000 UTC m=+51.419542055 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.275840 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.297537 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.313251 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:47Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.366099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.366155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.366173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.366197 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.366215 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.371197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.371437 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.371772 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs podName:ada6b953-f533-4b33-b07e-7e80604fe4a1 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:48.371743317 +0000 UTC m=+36.520784313 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs") pod "network-metrics-daemon-hjmwb" (UID: "ada6b953-f533-4b33-b07e-7e80604fe4a1") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.469081 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.469115 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.469124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.469137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.469147 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.572153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.572215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.572232 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.572259 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.572279 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.588336 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.588353 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.588421 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.588687 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.588970 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:47 crc kubenswrapper[4726]: E1123 20:08:47.589339 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.675043 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.675097 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.675114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.675137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.675157 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.778359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.778434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.778461 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.778492 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.778515 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.881019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.881752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.881773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.882680 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.882774 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.985728 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.985783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.985862 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.985910 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:47 crc kubenswrapper[4726]: I1123 20:08:47.985932 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:47Z","lastTransitionTime":"2025-11-23T20:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.088707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.088778 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.088796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.088821 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.088839 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.136085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.136140 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.136160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.136180 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.136196 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.155100 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:48Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.159616 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.159679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.159705 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.159740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.159767 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.180932 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:48Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.189178 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.189244 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.189261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.189287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.189308 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.212433 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:48Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.218039 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.218098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.218118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.218141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.218160 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.243735 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:48Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.249003 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.249270 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.249481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.249634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.249784 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.273415 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:48Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.274185 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.279107 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.279402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.279630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.279953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.280037 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.382930 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.383166 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.383278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.383372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.383466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.383316 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs podName:ada6b953-f533-4b33-b07e-7e80604fe4a1 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:50.383289021 +0000 UTC m=+38.532330017 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs") pod "network-metrics-daemon-hjmwb" (UID: "ada6b953-f533-4b33-b07e-7e80604fe4a1") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.383561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.383622 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.486957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.487222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.487347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.487469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.487565 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.588665 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:48 crc kubenswrapper[4726]: E1123 20:08:48.590022 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.590225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.590294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.590314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.590342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.590360 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.693270 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.693329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.693343 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.693369 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.693386 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.796441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.796482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.796492 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.796508 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.796518 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.899002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.899079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.899099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.899122 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:48 crc kubenswrapper[4726]: I1123 20:08:48.899141 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:48Z","lastTransitionTime":"2025-11-23T20:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.002524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.002600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.002625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.002654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.002677 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.105569 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.105735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.105754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.105777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.105794 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.209688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.209824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.209845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.209894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.209917 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.214002 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.237690 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.257539 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.279985 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.300744 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.314241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.314306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.314328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.314359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.314381 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.332346 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.351738 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.376149 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.395514 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.417580 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.417642 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.417661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.417690 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.417716 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.428722 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.449714 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.476170 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.494981 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.513197 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.520516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.520583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.520602 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.520627 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.520645 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.543020 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.556181 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.568192 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.581099 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:49Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.588683 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.588779 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:49 crc kubenswrapper[4726]: E1123 20:08:49.588852 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:49 crc kubenswrapper[4726]: E1123 20:08:49.588970 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.589086 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:49 crc kubenswrapper[4726]: E1123 20:08:49.589170 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.623414 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.623471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.623490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.623516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.623537 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.725980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.726023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.726036 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.726056 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.726068 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.828475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.828520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.828529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.828544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.828554 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.930631 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.930673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.930683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.930701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:49 crc kubenswrapper[4726]: I1123 20:08:49.930711 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:49Z","lastTransitionTime":"2025-11-23T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.033374 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.033699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.033833 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.033962 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.034061 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.136577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.136798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.136857 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.137052 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.137119 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.240477 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.240530 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.240547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.240571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.240588 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.343711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.343776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.343799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.343829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.343851 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.408141 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:50 crc kubenswrapper[4726]: E1123 20:08:50.408305 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:50 crc kubenswrapper[4726]: E1123 20:08:50.408740 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs podName:ada6b953-f533-4b33-b07e-7e80604fe4a1 nodeName:}" failed. No retries permitted until 2025-11-23 20:08:54.408710672 +0000 UTC m=+42.557751658 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs") pod "network-metrics-daemon-hjmwb" (UID: "ada6b953-f533-4b33-b07e-7e80604fe4a1") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.447523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.447903 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.448220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.448464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.448689 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.551156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.551446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.551654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.551917 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.552130 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.589130 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:50 crc kubenswrapper[4726]: E1123 20:08:50.589265 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.654738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.654808 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.654832 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.654862 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.654959 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.757617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.757827 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.757935 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.758008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.758067 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.861069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.861106 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.861115 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.861130 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.861139 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.963802 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.964171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.964200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.964223 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:50 crc kubenswrapper[4726]: I1123 20:08:50.964243 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:50Z","lastTransitionTime":"2025-11-23T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.067748 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.068023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.068150 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.068277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.068384 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.171707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.171967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.172056 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.172121 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.172183 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.274578 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.274632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.274649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.274673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.274689 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.377547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.377600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.377617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.377641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.377658 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.480080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.480361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.480480 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.480594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.480687 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.583225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.583271 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.583285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.583302 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.583314 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.588591 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.588635 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:51 crc kubenswrapper[4726]: E1123 20:08:51.588707 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.588846 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:51 crc kubenswrapper[4726]: E1123 20:08:51.589028 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:51 crc kubenswrapper[4726]: E1123 20:08:51.589204 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.690536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.690610 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.690637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.690668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.690693 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.794346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.794385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.794394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.794410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.794419 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.897973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.898032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.898049 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.898069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:51 crc kubenswrapper[4726]: I1123 20:08:51.898084 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:51Z","lastTransitionTime":"2025-11-23T20:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.000739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.000798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.000815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.000839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.000856 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.104055 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.104109 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.104124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.104143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.104158 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.206777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.206840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.206857 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.206919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.206957 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.310470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.310503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.310512 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.310543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.310553 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.413401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.413466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.413485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.413510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.413529 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.523542 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.523606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.523625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.523649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.523666 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.588368 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:52 crc kubenswrapper[4726]: E1123 20:08:52.588578 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.609741 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.626439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.626486 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.626502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.626527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.626544 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.626957 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.643975 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.661322 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.693169 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.713197 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.727626 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.730219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.730430 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.730569 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.730714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.730863 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.776850 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.801825 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.829400 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.834944 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.834977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.834985 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.834997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.835008 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.848905 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.860462 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.875132 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.889225 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.901214 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.913173 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.924152 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:52Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.937074 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.937122 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.937133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.937152 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:52 crc kubenswrapper[4726]: I1123 20:08:52.937164 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:52Z","lastTransitionTime":"2025-11-23T20:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.040329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.040615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.040699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.040798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.040944 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.144248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.144295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.144328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.144355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.144366 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.246839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.247182 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.247331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.247515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.247662 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.351239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.351609 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.351752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.351973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.352118 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.455489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.455562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.455580 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.455607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.455626 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.558518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.558587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.558605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.558627 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.558643 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.588076 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.588097 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.588238 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:53 crc kubenswrapper[4726]: E1123 20:08:53.588395 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:53 crc kubenswrapper[4726]: E1123 20:08:53.588769 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:53 crc kubenswrapper[4726]: E1123 20:08:53.588929 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.661601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.661643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.661659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.661682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.661701 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.764637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.764694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.764711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.764736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.764754 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.868455 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.868521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.868538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.868567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.868666 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.971921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.972325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.972355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.972381 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:53 crc kubenswrapper[4726]: I1123 20:08:53.972400 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:53Z","lastTransitionTime":"2025-11-23T20:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.075982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.076045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.076062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.076086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.076108 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.179621 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.179674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.179694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.179714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.179731 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.283192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.283611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.283826 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.284080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.284316 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.387726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.387778 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.387787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.387801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.387811 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.452832 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:54 crc kubenswrapper[4726]: E1123 20:08:54.453028 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:54 crc kubenswrapper[4726]: E1123 20:08:54.453119 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs podName:ada6b953-f533-4b33-b07e-7e80604fe4a1 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:02.453095225 +0000 UTC m=+50.602136201 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs") pod "network-metrics-daemon-hjmwb" (UID: "ada6b953-f533-4b33-b07e-7e80604fe4a1") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.491553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.491628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.491652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.491682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.491704 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.589035 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:54 crc kubenswrapper[4726]: E1123 20:08:54.589285 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.594309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.594381 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.594406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.594434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.594455 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.697025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.697066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.697077 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.697096 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.697108 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.800222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.800264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.800275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.800292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.800303 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.902632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.902704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.902732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.902763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:54 crc kubenswrapper[4726]: I1123 20:08:54.902787 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:54Z","lastTransitionTime":"2025-11-23T20:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.005603 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.005650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.005662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.005679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.005691 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.107851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.107964 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.107983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.108010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.108031 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.217918 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.217979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.217996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.218044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.218065 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.320387 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.320452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.320473 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.320502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.320522 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.423553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.423617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.423635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.423658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.423675 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.527040 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.527101 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.527118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.527142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.527161 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.588097 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.588132 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.588119 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:55 crc kubenswrapper[4726]: E1123 20:08:55.588281 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:55 crc kubenswrapper[4726]: E1123 20:08:55.588371 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:55 crc kubenswrapper[4726]: E1123 20:08:55.588479 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.630539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.630590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.630607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.630633 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.630652 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.733291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.733345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.733363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.733387 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.733403 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.835916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.835947 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.835957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.835971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.835981 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.938853 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.938942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.938959 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.938983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:55 crc kubenswrapper[4726]: I1123 20:08:55.939000 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:55Z","lastTransitionTime":"2025-11-23T20:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.041327 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.041399 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.041422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.041452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.041480 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.143606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.143682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.143705 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.143738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.143762 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.247466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.247501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.247509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.247523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.247533 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.350077 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.350110 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.350121 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.350137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.350149 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.452387 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.452454 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.452476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.452506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.452528 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.554611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.554668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.554685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.554738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.554756 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.588433 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:56 crc kubenswrapper[4726]: E1123 20:08:56.589009 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.657194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.657227 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.657255 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.657285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.657296 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.759925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.759992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.760010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.760035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.760052 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.862330 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.862380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.862394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.862412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.862425 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.965691 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.965755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.965775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.965801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:56 crc kubenswrapper[4726]: I1123 20:08:56.965821 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:56Z","lastTransitionTime":"2025-11-23T20:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.071372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.071460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.071518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.071546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.071563 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.174187 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.174228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.174246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.174266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.174282 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.277142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.277189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.277200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.277218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.277230 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.380008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.380084 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.380099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.380120 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.380135 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.483018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.483100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.483126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.483161 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.483187 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.587800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.587855 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.587887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.587905 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.587917 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.588156 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.588219 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.588256 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:57 crc kubenswrapper[4726]: E1123 20:08:57.588398 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:57 crc kubenswrapper[4726]: E1123 20:08:57.588615 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:57 crc kubenswrapper[4726]: E1123 20:08:57.588737 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.690357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.690403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.690418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.690433 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.690444 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.792916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.792974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.792995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.793018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.793036 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.895956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.896002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.896030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.896050 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.896064 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.998723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.998771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.998782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.998797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:57 crc kubenswrapper[4726]: I1123 20:08:57.998808 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:57Z","lastTransitionTime":"2025-11-23T20:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.102053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.102108 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.102125 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.102148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.102165 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.205903 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.205972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.205992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.206017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.206037 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.309125 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.309189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.309211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.309239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.309258 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.411791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.412180 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.412370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.412558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.412758 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.516148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.516231 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.516267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.516297 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.516318 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.568856 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.568942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.568967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.568995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.569016 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.588376 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:08:58 crc kubenswrapper[4726]: E1123 20:08:58.588672 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:08:58 crc kubenswrapper[4726]: E1123 20:08:58.595766 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:58Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.602452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.602509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.602527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.602552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.602572 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: E1123 20:08:58.622086 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:58Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.629313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.629446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.629470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.629506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.629544 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: E1123 20:08:58.657940 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:58Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.664040 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.664098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.664119 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.664146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.664165 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: E1123 20:08:58.684328 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:58Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.690508 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.690553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.690573 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.690596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.690616 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: E1123 20:08:58.711006 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:58Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:58 crc kubenswrapper[4726]: E1123 20:08:58.711230 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.713408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.713462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.713479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.713505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.713524 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.816920 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.816997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.817022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.817051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.817072 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.919685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.919731 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.919743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.919760 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:58 crc kubenswrapper[4726]: I1123 20:08:58.919772 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:58Z","lastTransitionTime":"2025-11-23T20:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.023859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.023952 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.023971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.023999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.024051 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.127346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.127405 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.127422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.127445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.127464 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.230441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.230511 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.230529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.230555 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.230572 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.333669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.333734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.333755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.333781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.333799 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.437337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.437402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.437419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.437444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.437463 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.540148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.540199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.540215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.540238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.540256 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.589017 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.589043 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.589062 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:08:59 crc kubenswrapper[4726]: E1123 20:08:59.589190 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:08:59 crc kubenswrapper[4726]: E1123 20:08:59.589338 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:08:59 crc kubenswrapper[4726]: E1123 20:08:59.591509 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.593925 4726 scope.go:117] "RemoveContainer" containerID="86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.643067 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.643133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.643153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.643179 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.643198 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.746362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.746427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.746451 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.746481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.746504 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.848979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.849027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.849044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.849064 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.849079 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.922455 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/1.log" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.926474 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.926946 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.939585 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:59Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.952325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.952351 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.952377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.952391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.952401 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:08:59Z","lastTransitionTime":"2025-11-23T20:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.955583 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:59Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.969519 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:59Z is after 2025-08-24T17:21:41Z" Nov 23 20:08:59 crc kubenswrapper[4726]: I1123 20:08:59.983052 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:08:59Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.003401 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.015994 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.038462 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.048941 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.054314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.054456 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.054560 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.054663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.054770 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.063481 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.081655 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.093821 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.109905 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.124298 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.152088 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.157673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.157707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.157716 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.157729 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.157739 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.166404 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.194661 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.210190 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.260303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.260366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.260385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.260407 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.260434 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.363043 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.363347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.363363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.363383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.363397 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.465175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.465225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.465234 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.465247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.465256 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.567254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.567289 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.567301 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.567316 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.567326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.588546 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:00 crc kubenswrapper[4726]: E1123 20:09:00.588666 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.669946 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.669976 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.669984 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.669999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.670008 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.771731 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.771766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.771774 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.771786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.771794 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.875201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.875294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.875321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.875362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.875388 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.933104 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/2.log" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.934039 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/1.log" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.937656 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259" exitCode=1 Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.937715 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.937772 4726 scope.go:117] "RemoveContainer" containerID="86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.939141 4726 scope.go:117] "RemoveContainer" containerID="7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259" Nov 23 20:09:00 crc kubenswrapper[4726]: E1123 20:09:00.939474 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.959967 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.980510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.980582 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.980602 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.980627 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.980646 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:00Z","lastTransitionTime":"2025-11-23T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.980979 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:00 crc kubenswrapper[4726]: I1123 20:09:00.992879 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:00Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.022845 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.041953 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.059976 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.076065 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.084933 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.084990 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.085007 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.085032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.085049 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.090354 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.104124 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.123095 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.141769 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.157944 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.171759 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.187825 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.187858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.187893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.187908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.187919 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.200893 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86abb8c780aef7e359a2fe5597582cb2e665bdc6ef4527f303dca80fc916b737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"message\\\":\\\"9lq openshift-kube-controller-manager/kube-controller-manager-crc openshift-network-node-identity/network-node-identity-vrzqb openshift-network-operator/iptables-alerter-4ln5h openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1123 20:08:42.697209 6037 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1123 20:08:42.697239 6037 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697264 6037 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697286 6037 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1123 20:08:42.697307 6037 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1123 20:08:42.697328 6037 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1123 20:08:42.697357 6037 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:08:42.697423 6037 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.221063 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.239643 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.254588 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.290475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.290539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.290559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.290584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.290601 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.394351 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.394414 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.394432 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.394459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.394478 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.498366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.498419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.498438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.498476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.498497 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.588900 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:01 crc kubenswrapper[4726]: E1123 20:09:01.589039 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.588921 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:01 crc kubenswrapper[4726]: E1123 20:09:01.589139 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.588908 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:01 crc kubenswrapper[4726]: E1123 20:09:01.589213 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.599960 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.599983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.599994 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.600007 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.600016 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.701395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.701644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.701709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.701770 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.701824 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.804092 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.804135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.804144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.804158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.804167 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.906543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.906584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.906599 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.906616 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.906626 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:01Z","lastTransitionTime":"2025-11-23T20:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.941809 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/2.log" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.945169 4726 scope.go:117] "RemoveContainer" containerID="7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259" Nov 23 20:09:01 crc kubenswrapper[4726]: E1123 20:09:01.945307 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.962740 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.977521 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:01 crc kubenswrapper[4726]: I1123 20:09:01.993014 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:01Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.008942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.008977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.008985 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.008997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.009006 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.011982 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.036162 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.053349 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.080073 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.097893 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.111858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.111926 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.111937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.111954 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.111968 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.114606 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.137017 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.154190 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.170943 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.200555 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.214193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.214239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.214272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.214296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.214314 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.214343 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.233276 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.250398 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.268831 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.316830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.316897 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.316907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.316924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.316935 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.419488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.419535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.419553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.419577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.419595 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.523201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.523545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.523707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.523941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.524137 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.544797 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:02 crc kubenswrapper[4726]: E1123 20:09:02.545085 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:09:02 crc kubenswrapper[4726]: E1123 20:09:02.545208 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs podName:ada6b953-f533-4b33-b07e-7e80604fe4a1 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:18.545179444 +0000 UTC m=+66.694220430 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs") pod "network-metrics-daemon-hjmwb" (UID: "ada6b953-f533-4b33-b07e-7e80604fe4a1") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.588518 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:02 crc kubenswrapper[4726]: E1123 20:09:02.588690 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.604113 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.628829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.628892 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.628901 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.628916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.628928 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.636008 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.665567 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.687704 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.706636 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.725689 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.730651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.730738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.730757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.730782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.730798 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.749279 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.764236 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.790564 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.813231 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.833165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.834647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.834713 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.834738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.834768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.834796 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.889926 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.902210 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.914028 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.931150 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.936624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.936645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.936654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.936665 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.936673 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:02Z","lastTransitionTime":"2025-11-23T20:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.948498 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:02 crc kubenswrapper[4726]: I1123 20:09:02.972778 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:02Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.039293 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.039332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.039340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.039354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.039363 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.142547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.143008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.143264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.143434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.143591 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.247011 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.247065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.247086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.247110 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.247126 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.289536 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.289752 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:09:35.289716186 +0000 UTC m=+83.438757182 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.290111 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.290351 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.290559 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.290743 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.290362 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.291110 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.291253 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.291455 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:35.291431643 +0000 UTC m=+83.440472639 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.290440 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.290639 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.290896 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.292024 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.292170 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.291779 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:35.291765682 +0000 UTC m=+83.440806638 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.292481 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:35.292460751 +0000 UTC m=+83.441501747 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.292621 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:35.292605555 +0000 UTC m=+83.441646541 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.351145 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.351205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.351222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.351245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.351262 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.454304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.454441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.454463 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.454503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.454519 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.557994 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.558176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.558203 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.558272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.558312 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.588413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.588492 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.588661 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.588862 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.589147 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:03 crc kubenswrapper[4726]: E1123 20:09:03.589390 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.662088 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.662394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.662646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.662796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.662965 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.766727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.766782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.766800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.766827 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.766844 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.869513 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.869567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.869586 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.869609 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.869625 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.972476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.972525 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.972542 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.972568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:03 crc kubenswrapper[4726]: I1123 20:09:03.972585 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:03Z","lastTransitionTime":"2025-11-23T20:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.075784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.075839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.075898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.075923 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.075940 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.179360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.179420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.179442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.179467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.179490 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.282889 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.282955 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.282973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.283000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.283018 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.386209 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.386266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.386280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.386299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.386316 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.489038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.489095 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.489114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.489137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.489155 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.588740 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:04 crc kubenswrapper[4726]: E1123 20:09:04.588911 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.591562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.591624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.591641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.591663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.591678 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.694371 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.694501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.694557 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.694584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.694635 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.797914 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.797999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.798025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.798059 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.798078 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.900648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.900696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.900708 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.900725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:04 crc kubenswrapper[4726]: I1123 20:09:04.900736 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:04Z","lastTransitionTime":"2025-11-23T20:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.003431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.003531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.003556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.003583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.003605 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.105738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.105816 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.105844 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.105912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.105940 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.209075 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.209120 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.209138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.209160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.209175 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.312391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.312439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.312455 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.312476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.312493 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.414779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.414916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.414932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.414949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.414961 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.517955 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.517989 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.517999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.518013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.518022 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.589118 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:05 crc kubenswrapper[4726]: E1123 20:09:05.589311 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.590135 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:05 crc kubenswrapper[4726]: E1123 20:09:05.590320 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.590142 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:05 crc kubenswrapper[4726]: E1123 20:09:05.590461 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.620571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.620624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.620645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.620674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.620695 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.723159 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.723199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.723210 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.723224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.723234 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.825211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.825254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.825265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.825283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.825293 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.928371 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.928481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.928542 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.928563 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:05 crc kubenswrapper[4726]: I1123 20:09:05.928578 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:05Z","lastTransitionTime":"2025-11-23T20:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.031189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.031243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.031263 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.031291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.031308 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.133275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.133310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.133318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.133332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.133342 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.236702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.236761 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.236784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.236811 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.236833 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.340651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.340701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.340718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.340741 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.340757 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.442552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.442611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.442628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.442647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.442661 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.545033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.545076 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.545089 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.545105 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.545119 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.588720 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:06 crc kubenswrapper[4726]: E1123 20:09:06.588859 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.647451 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.647504 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.647518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.647538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.647550 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.750407 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.750438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.750446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.750458 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.750466 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.853233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.853287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.853300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.853326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.853342 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.956129 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.956160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.956168 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.956235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:06 crc kubenswrapper[4726]: I1123 20:09:06.956248 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:06Z","lastTransitionTime":"2025-11-23T20:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.058685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.058735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.058743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.058754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.058766 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.161314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.161341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.161348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.161360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.161385 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.264712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.264742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.264753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.264774 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.264785 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.366943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.367001 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.367023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.367050 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.367072 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.470727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.470792 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.470809 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.470831 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.470848 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.573958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.574018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.574041 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.574068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.574115 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.588814 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.588860 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:07 crc kubenswrapper[4726]: E1123 20:09:07.588990 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.589093 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:07 crc kubenswrapper[4726]: E1123 20:09:07.589185 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:07 crc kubenswrapper[4726]: E1123 20:09:07.589374 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.677083 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.677112 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.677120 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.677133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.677142 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.779951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.780362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.780385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.780413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.780431 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.883057 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.883124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.883147 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.883174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:07 crc kubenswrapper[4726]: I1123 20:09:07.883195 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:07Z","lastTransitionTime":"2025-11-23T20:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.022066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.022134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.022157 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.022184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.022206 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.027050 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.035230 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.043816 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.075188 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.095129 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.115628 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.124845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.124943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.124970 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.124999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.125020 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.132709 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.147348 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.168361 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.188924 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.207246 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.227228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.227298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.227324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.227359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.227316 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.227383 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.245297 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.266014 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.282382 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.292235 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.303117 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.313721 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.330403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.330453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.330472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.330495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.330515 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.332400 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.432797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.432903 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.432921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.432945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.432963 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.534782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.534851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.534891 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.534926 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.534944 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.591573 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:08 crc kubenswrapper[4726]: E1123 20:09:08.591704 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.637256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.637333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.637349 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.637374 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.637391 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.739967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.740032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.740054 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.740080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.740101 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.843493 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.843545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.843559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.843574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.843585 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.856468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.856551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.856568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.856593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.856611 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: E1123 20:09:08.869267 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.873299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.873366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.873388 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.873416 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.873440 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: E1123 20:09:08.892082 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.895827 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.895911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.895935 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.895962 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.895990 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: E1123 20:09:08.913335 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.917359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.917406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.917422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.917476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.917493 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: E1123 20:09:08.938851 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.946035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.946075 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.946086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.946102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.946113 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:08 crc kubenswrapper[4726]: E1123 20:09:08.970979 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:08Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:08 crc kubenswrapper[4726]: E1123 20:09:08.971228 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.972980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.973017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.973033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.973055 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:08 crc kubenswrapper[4726]: I1123 20:09:08.973072 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:08Z","lastTransitionTime":"2025-11-23T20:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.074571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.074617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.074629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.074644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.074981 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.177281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.177311 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.177322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.177336 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.177346 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.279383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.279408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.279416 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.279428 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.279436 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.381198 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.381229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.381240 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.381256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.381268 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.484272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.484323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.484340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.484362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.484378 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.587005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.587047 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.587058 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.587076 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.587087 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.588185 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:09 crc kubenswrapper[4726]: E1123 20:09:09.588293 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.588343 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:09 crc kubenswrapper[4726]: E1123 20:09:09.588394 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.588435 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:09 crc kubenswrapper[4726]: E1123 20:09:09.588487 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.689490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.689524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.689537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.689553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.689564 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.791453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.791483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.791494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.791510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.791521 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.894794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.895137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.895151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.895172 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.895185 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.998175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.998263 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.998288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.998318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:09 crc kubenswrapper[4726]: I1123 20:09:09.998341 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:09Z","lastTransitionTime":"2025-11-23T20:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.100943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.101006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.101025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.101051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.101068 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.203552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.203632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.203659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.203688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.203711 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.306558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.306629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.306653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.306755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.306783 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.410308 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.410379 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.410403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.410430 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.410463 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.513979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.514053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.514112 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.514144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.514168 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.589225 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:10 crc kubenswrapper[4726]: E1123 20:09:10.589703 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.616556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.616604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.616621 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.616641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.616658 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.720013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.720064 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.720081 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.720105 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.720122 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.823119 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.823189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.823213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.823243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.823266 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.925900 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.926151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.926240 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.926347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:10 crc kubenswrapper[4726]: I1123 20:09:10.926469 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:10Z","lastTransitionTime":"2025-11-23T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.029496 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.029545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.029563 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.029588 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.029606 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.132847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.132964 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.132991 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.133023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.133052 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.236647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.236709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.236730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.236761 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.236786 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.340146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.340213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.340236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.340262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.340282 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.442930 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.442983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.443001 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.443027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.443047 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.546712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.546751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.546759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.546775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.546784 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.588248 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:11 crc kubenswrapper[4726]: E1123 20:09:11.588417 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.588675 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:11 crc kubenswrapper[4726]: E1123 20:09:11.588775 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.589004 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:11 crc kubenswrapper[4726]: E1123 20:09:11.589098 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.649256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.649294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.649304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.649321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.649331 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.751155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.751261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.751304 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.751321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.751336 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.854003 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.854044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.854056 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.854074 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.854086 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.956394 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.956468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.956490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.956518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:11 crc kubenswrapper[4726]: I1123 20:09:11.956543 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:11Z","lastTransitionTime":"2025-11-23T20:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.059104 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.059170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.059192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.059216 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.059233 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.162536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.162600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.162619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.162646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.162669 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.265379 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.265420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.265435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.265454 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.265468 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.368744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.368815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.368839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.368912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.368940 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.472126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.472204 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.472228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.472256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.472279 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.578581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.578657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.578682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.578715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.578741 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.588146 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:12 crc kubenswrapper[4726]: E1123 20:09:12.588378 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.612350 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.630740 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.651906 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.671181 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.680456 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.680489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.680500 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.680536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.680547 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.699373 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.717218 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.737480 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.753724 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.769947 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.783174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.783207 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.783218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.783234 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.783246 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.784674 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.805557 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.816655 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.839701 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.855611 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5118a70b-3f3c-483d-9e77-bb83969ebcf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.873365 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.885480 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.885553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.885606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.885629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.885659 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.887252 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.899265 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.912025 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:12Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.993724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.993788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.993988 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.994009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:12 crc kubenswrapper[4726]: I1123 20:09:12.994020 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:12Z","lastTransitionTime":"2025-11-23T20:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.096506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.096922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.097129 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.097404 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.097565 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.200967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.201030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.201053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.201080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.201106 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.304673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.304714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.304729 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.304752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.304764 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.408099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.408396 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.408549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.408804 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.409040 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.512622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.512676 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.512695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.512720 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.512742 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.588520 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:13 crc kubenswrapper[4726]: E1123 20:09:13.588677 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.588988 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:13 crc kubenswrapper[4726]: E1123 20:09:13.589108 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.589330 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:13 crc kubenswrapper[4726]: E1123 20:09:13.589424 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.615468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.615510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.615529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.615552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.615569 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.718662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.718724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.718740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.718766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.718784 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.822269 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.822509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.822571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.822641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.822702 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.925473 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.925530 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.925548 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.925574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:13 crc kubenswrapper[4726]: I1123 20:09:13.925591 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:13Z","lastTransitionTime":"2025-11-23T20:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.028028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.028098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.028123 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.028154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.028177 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.131344 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.131414 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.131432 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.131463 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.131488 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.235998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.236091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.236114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.236144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.236164 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.340242 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.340306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.340322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.340348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.340365 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.448040 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.448102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.448127 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.448222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.448245 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.551536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.551602 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.551627 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.551658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.551687 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.589315 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:14 crc kubenswrapper[4726]: E1123 20:09:14.589513 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.590730 4726 scope.go:117] "RemoveContainer" containerID="7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259" Nov 23 20:09:14 crc kubenswrapper[4726]: E1123 20:09:14.591003 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.654253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.654329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.654356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.654380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.654397 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.756679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.756751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.756770 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.756793 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.756810 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.859518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.859559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.859571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.859585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.859594 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.962705 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.962766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.962794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.962823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:14 crc kubenswrapper[4726]: I1123 20:09:14.962847 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:14Z","lastTransitionTime":"2025-11-23T20:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.064788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.064824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.064835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.064850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.064862 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.167012 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.167051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.167062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.167079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.167090 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.269329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.269378 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.269398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.269422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.269440 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.372475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.372532 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.372550 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.372572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.372590 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.475669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.475742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.475760 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.475787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.475805 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.578758 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.578802 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.578811 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.578824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.578833 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.588505 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.588570 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.588502 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:15 crc kubenswrapper[4726]: E1123 20:09:15.588663 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:15 crc kubenswrapper[4726]: E1123 20:09:15.588971 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:15 crc kubenswrapper[4726]: E1123 20:09:15.589156 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.681707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.681790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.681815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.681847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.681906 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.784053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.784118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.784136 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.784159 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.784176 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.886594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.886666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.886683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.886712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.886738 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.988864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.988958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.988998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.989023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:15 crc kubenswrapper[4726]: I1123 20:09:15.989040 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:15Z","lastTransitionTime":"2025-11-23T20:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.091751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.091797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.091806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.091822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.091832 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.194482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.194538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.194555 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.194581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.194599 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.297989 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.298058 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.298081 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.298110 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.298130 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.400262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.400291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.400299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.400333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.400343 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.502280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.502311 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.502319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.502331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.502340 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.588463 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:16 crc kubenswrapper[4726]: E1123 20:09:16.588640 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.604683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.604718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.604730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.604748 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.604760 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.707743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.707788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.707800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.707817 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.707829 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.809843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.809899 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.809911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.809931 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.809942 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.912473 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.912533 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.912552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.912577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:16 crc kubenswrapper[4726]: I1123 20:09:16.912596 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:16Z","lastTransitionTime":"2025-11-23T20:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.015408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.015470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.015487 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.015511 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.015530 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.118045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.118078 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.118114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.118127 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.118136 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.219700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.219744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.219754 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.219769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.219779 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.322417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.322461 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.322470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.322485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.322495 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.424364 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.424414 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.424430 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.424456 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.424468 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.527105 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.527138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.527147 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.527160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.527170 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.588932 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.589023 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.588971 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:17 crc kubenswrapper[4726]: E1123 20:09:17.589102 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:17 crc kubenswrapper[4726]: E1123 20:09:17.589244 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:17 crc kubenswrapper[4726]: E1123 20:09:17.589304 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.629205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.629252 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.629262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.629275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.629284 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.730957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.730990 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.730998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.731013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.731023 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.832723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.832750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.832757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.832769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.832778 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.935110 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.935154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.935166 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.935182 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:17 crc kubenswrapper[4726]: I1123 20:09:17.935193 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:17Z","lastTransitionTime":"2025-11-23T20:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.037207 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.037245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.037254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.037267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.037276 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.140322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.140363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.140372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.140389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.140399 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.242929 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.243013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.243031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.243086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.243107 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.345380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.345419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.345431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.345447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.345461 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.447927 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.447977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.447988 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.448002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.448012 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.550786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.550849 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.550899 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.550928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.550950 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.588368 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:18 crc kubenswrapper[4726]: E1123 20:09:18.588669 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.624169 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:18 crc kubenswrapper[4726]: E1123 20:09:18.624277 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:09:18 crc kubenswrapper[4726]: E1123 20:09:18.624323 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs podName:ada6b953-f533-4b33-b07e-7e80604fe4a1 nodeName:}" failed. No retries permitted until 2025-11-23 20:09:50.624310451 +0000 UTC m=+98.773351407 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs") pod "network-metrics-daemon-hjmwb" (UID: "ada6b953-f533-4b33-b07e-7e80604fe4a1") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.653526 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.653712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.653727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.653741 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.653753 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.756082 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.756318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.756418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.756514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.756756 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.859107 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.859135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.859143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.859156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.859164 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.961278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.961481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.961583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.961668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:18 crc kubenswrapper[4726]: I1123 20:09:18.961744 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:18Z","lastTransitionTime":"2025-11-23T20:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.064236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.064280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.064289 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.064305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.064316 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.167098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.167168 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.167185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.167208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.167227 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.256669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.256710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.256722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.256738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.256750 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.275398 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:19Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.279577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.279620 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.279632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.279646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.279709 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.293348 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:19Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.296928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.297099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.297160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.297241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.297307 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.315942 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:19Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.319309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.319338 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.319348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.319363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.319372 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.336734 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:19Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.340124 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.340175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.340190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.340211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.340225 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.355152 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:19Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.355290 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.356618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.356661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.356671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.356685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.356694 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.459184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.459228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.459241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.459261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.459276 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.561544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.561573 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.561581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.561593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.561601 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.588429 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.588437 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.588548 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.588713 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.588766 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:19 crc kubenswrapper[4726]: E1123 20:09:19.588826 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.663990 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.664047 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.664064 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.664087 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.664105 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.766165 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.766226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.766238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.766255 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.766639 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.869764 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.869788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.869795 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.869807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.869823 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.971392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.971658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.971728 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.971797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:19 crc kubenswrapper[4726]: I1123 20:09:19.971855 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:19Z","lastTransitionTime":"2025-11-23T20:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.074327 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.074655 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.074787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.074943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.075042 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.178629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.179073 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.179440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.179674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.180001 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.283286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.283334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.283345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.283364 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.283381 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.386541 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.386612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.386638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.386668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.386690 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.489684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.489725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.489736 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.489750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.489759 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.588444 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:20 crc kubenswrapper[4726]: E1123 20:09:20.588660 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.592492 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.592717 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.592948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.593141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.593318 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.695834 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.695895 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.695907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.695923 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.695935 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.799018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.799047 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.799055 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.799067 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.799075 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.901278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.901669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.901972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.902177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:20 crc kubenswrapper[4726]: I1123 20:09:20.902357 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:20Z","lastTransitionTime":"2025-11-23T20:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.006606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.006648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.006657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.006677 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.006686 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.061004 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/0.log" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.061079 4726 generic.go:334] "Generic (PLEG): container finished" podID="8c5753d1-1307-4d28-b08f-3588c14ec1bd" containerID="0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec" exitCode=1 Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.061123 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jttr8" event={"ID":"8c5753d1-1307-4d28-b08f-3588c14ec1bd","Type":"ContainerDied","Data":"0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.061657 4726 scope.go:117] "RemoveContainer" containerID="0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.087137 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.103738 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.114107 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.114156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.114173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.114196 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.114214 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.139756 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.153617 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5118a70b-3f3c-483d-9e77-bb83969ebcf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.169379 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.188278 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.207671 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.217938 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.217979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.217995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.218017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.218033 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.232330 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.244855 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.255581 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.271594 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.284705 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.296655 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:20Z\\\",\\\"message\\\":\\\"2025-11-23T20:08:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76\\\\n2025-11-23T20:08:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76 to /host/opt/cni/bin/\\\\n2025-11-23T20:08:35Z [verbose] multus-daemon started\\\\n2025-11-23T20:08:35Z [verbose] Readiness Indicator file check\\\\n2025-11-23T20:09:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.308921 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.320302 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.320350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.320366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.320389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.320406 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.327030 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.339576 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.352179 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.361693 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:21Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.422516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.422545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.422554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.422565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.422573 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.524577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.524606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.524614 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.524627 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.524635 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.588205 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:21 crc kubenswrapper[4726]: E1123 20:09:21.588293 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.588407 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:21 crc kubenswrapper[4726]: E1123 20:09:21.588448 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.588539 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:21 crc kubenswrapper[4726]: E1123 20:09:21.588582 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.626789 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.626832 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.626849 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.626901 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.626919 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.729287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.729319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.729328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.729340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.729349 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.831932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.831982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.831991 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.832002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.832012 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.934577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.934620 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.934629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.934650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:21 crc kubenswrapper[4726]: I1123 20:09:21.934659 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:21Z","lastTransitionTime":"2025-11-23T20:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.038384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.038444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.038462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.038484 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.038501 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.067474 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/0.log" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.067611 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jttr8" event={"ID":"8c5753d1-1307-4d28-b08f-3588c14ec1bd","Type":"ContainerStarted","Data":"268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.088281 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.105470 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.123976 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:20Z\\\",\\\"message\\\":\\\"2025-11-23T20:08:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76\\\\n2025-11-23T20:08:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76 to /host/opt/cni/bin/\\\\n2025-11-23T20:08:35Z [verbose] multus-daemon started\\\\n2025-11-23T20:08:35Z [verbose] Readiness Indicator file check\\\\n2025-11-23T20:09:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.139742 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.141235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.141277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.141287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.141301 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.141310 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.170180 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.188893 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.208414 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.220856 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.229435 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.244298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.244336 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.244347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.244363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.244376 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.247449 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.261419 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5118a70b-3f3c-483d-9e77-bb83969ebcf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.277110 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.288242 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.299594 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.313543 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.329095 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.344683 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.347850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.347922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.347939 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.347967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.347989 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.361097 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.450515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.450712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.450849 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.451040 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.451167 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.553486 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.554356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.554519 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.554664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.554790 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.588609 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:22 crc kubenswrapper[4726]: E1123 20:09:22.588696 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.606632 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.622119 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.637754 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.650328 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.659291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.659342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.659353 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.659368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.659379 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.674034 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.684348 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.704306 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.718784 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5118a70b-3f3c-483d-9e77-bb83969ebcf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.729943 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.743026 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.756354 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.762054 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.762224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.762318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.762420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.762502 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.772210 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:20Z\\\",\\\"message\\\":\\\"2025-11-23T20:08:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76\\\\n2025-11-23T20:08:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76 to /host/opt/cni/bin/\\\\n2025-11-23T20:08:35Z [verbose] multus-daemon started\\\\n2025-11-23T20:08:35Z [verbose] Readiness Indicator file check\\\\n2025-11-23T20:09:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.786915 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.800712 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.819305 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.829711 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.843960 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.859286 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:22Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.864931 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.864986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.865003 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.865029 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.865045 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.967047 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.967358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.967481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.967585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:22 crc kubenswrapper[4726]: I1123 20:09:22.967716 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:22Z","lastTransitionTime":"2025-11-23T20:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.071277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.071332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.071346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.071363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.071374 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.173378 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.173407 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.173417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.173430 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.173438 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.277087 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.277134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.277145 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.277160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.277168 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.381121 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.381203 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.381224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.381253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.381282 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.485933 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.485970 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.485979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.485992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.486002 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.588200 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.588269 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.588212 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.588330 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.588348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.588372 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: E1123 20:09:23.588291 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:23 crc kubenswrapper[4726]: E1123 20:09:23.588394 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.588389 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.588912 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:23 crc kubenswrapper[4726]: E1123 20:09:23.588969 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.692682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.692732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.692749 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.692771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.692790 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.795299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.795326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.795334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.795347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.795357 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.896820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.896906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.896928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.896950 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.896968 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.999349 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.999395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.999412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.999433 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:23 crc kubenswrapper[4726]: I1123 20:09:23.999450 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:23Z","lastTransitionTime":"2025-11-23T20:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.100592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.100617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.100626 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.100637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.100645 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.203153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.203177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.203186 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.203199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.203208 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.306077 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.306111 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.306122 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.306134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.306143 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.408495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.408529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.408537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.408549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.408558 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.510861 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.510919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.510932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.510948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.510959 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.588679 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:24 crc kubenswrapper[4726]: E1123 20:09:24.588767 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.612467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.612546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.612562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.612577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.612610 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.714535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.714600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.714617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.714639 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.714655 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.816282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.816325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.816335 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.816350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.816359 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.918596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.918643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.918653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.918670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:24 crc kubenswrapper[4726]: I1123 20:09:24.918678 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:24Z","lastTransitionTime":"2025-11-23T20:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.021267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.021306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.021318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.021334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.021343 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.123928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.123982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.123993 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.124008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.124017 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.226542 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.226579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.226588 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.226602 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.226611 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.329237 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.329274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.329283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.329296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.329306 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.431298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.431345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.431356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.431370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.431395 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.532930 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.532968 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.532977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.532990 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.533000 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.588582 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.588618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:25 crc kubenswrapper[4726]: E1123 20:09:25.588689 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.588624 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:25 crc kubenswrapper[4726]: E1123 20:09:25.588785 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:25 crc kubenswrapper[4726]: E1123 20:09:25.589020 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.635527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.635590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.635598 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.635610 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.635619 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.737981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.738008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.738016 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.738028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.738037 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.840375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.840410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.840421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.840435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.840446 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.942916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.942977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.943002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.943030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:25 crc kubenswrapper[4726]: I1123 20:09:25.943051 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:25Z","lastTransitionTime":"2025-11-23T20:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.045230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.045303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.045320 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.045345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.045363 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.148487 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.148528 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.148538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.148554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.148566 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.250795 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.250848 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.250876 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.250893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.250927 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.353597 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.353649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.353666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.353691 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.353708 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.456913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.457028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.457042 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.457350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.457366 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.560174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.560244 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.560266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.560295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.560317 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.588804 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:26 crc kubenswrapper[4726]: E1123 20:09:26.589028 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.603402 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.663176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.663239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.663258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.663282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.663302 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.765977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.766011 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.766022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.766038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.766049 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.869772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.869812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.869828 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.869851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.869890 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.972919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.972962 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.972978 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.972999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:26 crc kubenswrapper[4726]: I1123 20:09:26.973016 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:26Z","lastTransitionTime":"2025-11-23T20:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.076070 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.076114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.076130 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.076151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.076167 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.179118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.179183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.179202 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.179226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.179245 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.283083 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.283171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.283190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.283617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.283676 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.386266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.386354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.386373 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.386432 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.386455 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.489100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.489149 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.489167 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.489190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.489210 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.588299 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.588367 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.588302 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:27 crc kubenswrapper[4726]: E1123 20:09:27.588466 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:27 crc kubenswrapper[4726]: E1123 20:09:27.588640 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:27 crc kubenswrapper[4726]: E1123 20:09:27.588756 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.591755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.591798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.591815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.591836 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.591854 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.694830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.694908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.694927 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.694950 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.694967 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.797850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.797994 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.798022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.798055 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.798079 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.902046 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.902174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.902236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.902261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:27 crc kubenswrapper[4726]: I1123 20:09:27.902283 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:27Z","lastTransitionTime":"2025-11-23T20:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.006128 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.006195 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.006212 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.006239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.006257 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.109532 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.109606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.109629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.109660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.109683 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.212424 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.212474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.212491 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.212513 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.212530 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.316147 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.316205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.316223 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.316245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.316262 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.425922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.425997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.426017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.426043 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.426061 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.528462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.528527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.528548 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.528572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.528589 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.588473 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:28 crc kubenswrapper[4726]: E1123 20:09:28.588669 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.632066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.632177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.632199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.632224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.632242 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.735171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.735233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.735270 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.735292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.735310 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.838185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.838241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.838258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.838283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.838300 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.940863 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.940950 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.940973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.940999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:28 crc kubenswrapper[4726]: I1123 20:09:28.941022 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:28Z","lastTransitionTime":"2025-11-23T20:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.043697 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.043762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.043782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.043803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.043821 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.146772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.146818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.146834 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.146855 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.146897 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.249911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.249960 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.249974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.249992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.250005 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.353087 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.353159 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.353171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.353188 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.353199 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.456099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.456169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.456189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.456214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.456234 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.558434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.558499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.558519 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.558544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.558583 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.588422 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.588469 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.588617 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.588920 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.589036 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.589266 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.590358 4726 scope.go:117] "RemoveContainer" containerID="7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.661393 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.661447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.661465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.661490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.661508 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.718382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.718428 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.718444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.718466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.718482 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.738659 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:29Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.743352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.743403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.743421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.743443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.743460 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.764681 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:29Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.769606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.769645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.769663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.769685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.769702 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.793189 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:29Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.798222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.798268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.798286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.798309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.798326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.817230 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:29Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.827730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.827770 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.827782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.827800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.827815 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.846316 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"998bc227-8b31-4e93-8b90-99ca7f2d4cd9\\\",\\\"systemUUID\\\":\\\"d521adaa-c97b-4677-a80e-46a3d7f7a33d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:29Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:29 crc kubenswrapper[4726]: E1123 20:09:29.846836 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.849101 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.849143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.849158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.849175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.849188 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.951835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.951901 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.951920 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.951942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:29 crc kubenswrapper[4726]: I1123 20:09:29.951958 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:29Z","lastTransitionTime":"2025-11-23T20:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.055038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.055410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.055427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.055448 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.055545 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.095005 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/2.log" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.098141 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.099384 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.155185 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.158416 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.158463 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.158478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.158497 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.158512 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.181914 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.193033 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.206644 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.219023 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.230373 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.243493 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.253204 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.261005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.261048 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.261059 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.261078 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.261089 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.275801 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.287127 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5118a70b-3f3c-483d-9e77-bb83969ebcf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.297828 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.308621 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.317799 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.327646 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:20Z\\\",\\\"message\\\":\\\"2025-11-23T20:08:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76\\\\n2025-11-23T20:08:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76 to /host/opt/cni/bin/\\\\n2025-11-23T20:08:35Z [verbose] multus-daemon started\\\\n2025-11-23T20:08:35Z [verbose] Readiness Indicator file check\\\\n2025-11-23T20:09:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.342165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c1453c-20f8-4d3a-826b-4c80ed43d1ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0025d40e9e0a62c038a83d3258a2f3e72c4e33423152f7ee0f9e53430e22e23b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.357521 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.363160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.363202 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.363214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.363229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.363240 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.372371 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.391424 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.418362 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:09:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:30Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.465894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.465928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.465936 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.465950 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.465959 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.568375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.568412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.568423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.568440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.568449 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.588157 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:30 crc kubenswrapper[4726]: E1123 20:09:30.588274 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.672225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.672279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.672299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.672322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.672339 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.775228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.775471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.775538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.775633 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.775704 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.877756 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.877790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.877803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.877818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.877828 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.980458 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.980527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.980567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.980591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:30 crc kubenswrapper[4726]: I1123 20:09:30.980607 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:30Z","lastTransitionTime":"2025-11-23T20:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.083326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.083376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.083388 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.083405 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.083416 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.104772 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/3.log" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.105999 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/2.log" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.110301 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" exitCode=1 Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.110341 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.110376 4726 scope.go:117] "RemoveContainer" containerID="7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.120623 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:09:31 crc kubenswrapper[4726]: E1123 20:09:31.121197 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.146829 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.167289 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.188211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.188241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.188249 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.188261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.188271 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.190203 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.207646 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.231567 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.243972 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5118a70b-3f3c-483d-9e77-bb83969ebcf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.255375 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.269195 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.280004 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.288435 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.290505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.290546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.290559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.290660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.290676 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.301344 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c1453c-20f8-4d3a-826b-4c80ed43d1ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0025d40e9e0a62c038a83d3258a2f3e72c4e33423152f7ee0f9e53430e22e23b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.323001 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.336159 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.348906 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:20Z\\\",\\\"message\\\":\\\"2025-11-23T20:08:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76\\\\n2025-11-23T20:08:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76 to /host/opt/cni/bin/\\\\n2025-11-23T20:08:35Z [verbose] multus-daemon started\\\\n2025-11-23T20:08:35Z [verbose] Readiness Indicator file check\\\\n2025-11-23T20:09:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.360430 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.382760 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5250cd2353e3862ba32b266358798c125bffd190b35aa2910e2997a8a58259\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:00Z\\\",\\\"message\\\":\\\"rvice_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1123 20:09:00.564789 6257 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1123 20:09:00.564372 6257 obj_retry.go:303] R\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:30Z\\\",\\\"message\\\":\\\"oMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.253:2379: 10.217.5.253:9979:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1123 20:09:30.538626 6613 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538633 6613 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538643 6613 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1123 20:09:30.538648 6613 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1123 20:09:30.538652 6613 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538663 6613 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:09:30.538705 6613 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:09:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.392762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.392806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.392823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.392846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.392882 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.402322 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.421357 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.435773 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:31Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.495454 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.495496 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.495507 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.495527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.495539 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.588085 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.588143 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.588136 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:31 crc kubenswrapper[4726]: E1123 20:09:31.588292 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:31 crc kubenswrapper[4726]: E1123 20:09:31.588595 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:31 crc kubenswrapper[4726]: E1123 20:09:31.588827 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.598171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.598206 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.598221 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.598236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.598247 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.701233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.701296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.701316 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.701339 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.701356 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.803797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.803884 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.803902 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.803926 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.803944 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.908110 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.908171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.908190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.908214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:31 crc kubenswrapper[4726]: I1123 20:09:31.908237 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:31Z","lastTransitionTime":"2025-11-23T20:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.009793 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.009841 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.009850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.009881 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.009900 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.112127 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.112155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.112163 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.112176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.112186 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.114235 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/3.log" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.116918 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:09:32 crc kubenswrapper[4726]: E1123 20:09:32.117041 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.137480 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:30Z\\\",\\\"message\\\":\\\"oMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.253:2379: 10.217.5.253:9979:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1123 20:09:30.538626 6613 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538633 6613 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538643 6613 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1123 20:09:30.538648 6613 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1123 20:09:30.538652 6613 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538663 6613 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:09:30.538705 6613 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:09:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.147933 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.162923 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.173990 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.190460 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.202996 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5118a70b-3f3c-483d-9e77-bb83969ebcf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.233002 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.247676 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.257155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.257188 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.257199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.257214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.257225 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.267005 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.281439 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.293772 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.304673 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.329764 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.341128 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.351316 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.359828 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.360033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.360127 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.360195 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.360254 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.364406 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.375285 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.390635 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:20Z\\\",\\\"message\\\":\\\"2025-11-23T20:08:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76\\\\n2025-11-23T20:08:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76 to /host/opt/cni/bin/\\\\n2025-11-23T20:08:35Z [verbose] multus-daemon started\\\\n2025-11-23T20:08:35Z [verbose] Readiness Indicator file check\\\\n2025-11-23T20:09:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.402316 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c1453c-20f8-4d3a-826b-4c80ed43d1ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0025d40e9e0a62c038a83d3258a2f3e72c4e33423152f7ee0f9e53430e22e23b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.462280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.462314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.462323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.462336 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.462346 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.564396 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.564444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.564459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.564479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.564494 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.588161 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:32 crc kubenswrapper[4726]: E1123 20:09:32.588380 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.599169 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c1453c-20f8-4d3a-826b-4c80ed43d1ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0025d40e9e0a62c038a83d3258a2f3e72c4e33423152f7ee0f9e53430e22e23b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69530c12a167fc7f27d5e18f1de7bed9193df165e1ed16327a5cb980750977d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.613324 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5f61eba-3d7d-4a48-b943-27b509795c28\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3181aae2d33a96bb548fb3d53b1c4984c183c771d5f305d1b8eb7e53f2f84d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4c35be1c8e2e7049630ad8bf13f20ee7e560dbf237f5b43fd5a2a8e301d6d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95fa962759c2e5f089e172983cacc55c7e1902cd82b51d7f28508b616efd7be1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86899af9add2fde3593f2b320589c5d6a6cbe8e440e849a00380ee7e836076c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9beda7b1205e6b900aa7a00fa2a2e2e811b5ca251c0d8a5e582c2f22aeab6af3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"heck-endpoints-signer@1763928495\\\\\\\\\\\\\\\" (2025-11-23 20:08:15 +0000 UTC to 2025-12-23 20:08:16 +0000 UTC (now=2025-11-23 20:08:31.474888461 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475051 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475091 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1763928496\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1763928496\\\\\\\\\\\\\\\" (2025-11-23 19:08:16 +0000 UTC to 2026-11-23 19:08:16 +0000 UTC (now=2025-11-23 20:08:31.475065816 +0000 UTC))\\\\\\\"\\\\nI1123 20:08:31.475111 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1123 20:08:31.475121 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1123 20:08:31.475137 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1123 20:08:31.475195 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475218 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1123 20:08:31.475238 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1123 20:08:31.475258 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1123 20:08:31.475377 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1123 20:08:31.475390 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1123 20:08:31.476117 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1123 20:08:31.481233 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a933f2917cb0180379e795fccb04a05ff1957cdeadca43cd8dcc8c02a9d927b5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedf467814148c054081ce35a9ecebbc67a765b044dcfc718e9922a28c5c4bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.627208 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e23ba1b62a2bdd9c4a7df614cf162a388a846b4f7c95d5fc4ae1ca3128c1f5b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.639888 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jttr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c5753d1-1307-4d28-b08f-3588c14ec1bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:20Z\\\",\\\"message\\\":\\\"2025-11-23T20:08:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76\\\\n2025-11-23T20:08:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_87cb8dc5-b67c-4d7b-954e-e6bdda934b76 to /host/opt/cni/bin/\\\\n2025-11-23T20:08:35Z [verbose] multus-daemon started\\\\n2025-11-23T20:08:35Z [verbose] Readiness Indicator file check\\\\n2025-11-23T20:09:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-474m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jttr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.648295 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bq9lq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c99d511-e691-4e43-a4cd-fbd23a033b92\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebea75bf0cda5b4b5a92efb0385c22250fb7d58a6b175e39e7af67d2d6cb5acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfz7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bq9lq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.666084 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.666111 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.666120 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.666134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.666144 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.666905 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-23T20:09:30Z\\\",\\\"message\\\":\\\"oMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.253:2379: 10.217.5.253:9979:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {de17f0de-cfb1-4534-bb42-c40f5e050c73}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1123 20:09:30.538626 6613 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538633 6613 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538643 6613 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1123 20:09:30.538648 6613 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc after 0 failed attempt(s)\\\\nI1123 20:09:30.538652 6613 default_network_controller.go:776] Recording success event on pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1123 20:09:30.538663 6613 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1123 20:09:30.538705 6613 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-23T20:09:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf84x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-td7h8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.682581 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ec796ad-507c-4d48-bc55-7f23419d4d98\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af64b0e5d391f66cb2fd00996f1ed23fbd4c725c703a315e32806c420be7563\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967057ee97e54080cfdde9a43a50a14ef5f99cb460c8ce22724c2441e27eed4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d8c3e7966e95966d51403ea87df9429c44dd581c916286054088723a6a21422\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://661f6260e97c7b78210fa07a882551ca3e3b3a5b49928b6b86b44ba50b586e24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.694665 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b743bda623f06c105dbab5e8327dd44a2550947e4ed3409e5d765bd6aa906b82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.708139 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e3ac186-9f76-4774-8e04-fb00add1eb72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://014f154b5686a69643b12d92e3b50867807febf68dcb57702e9ebc68369f6b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tcgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c58qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.724968 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.740055 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a041870-99e1-48bc-99a6-f1c3c01fa5ce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fe6b9eeb9c4ac45d44f800b8291b36fa4091ee26d9540d4071ae084ae4f32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d118cef46cdde57a7a0b45c95ffa85f0ab033bcc188dfb7e702e21e66482ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5846491dc11debfc2e37b6a5ed1ddabb55da64f23ab9946bea70f992f2369a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7f7b5d51c12042b51227cebf269721949bbfd22d23f0bf9f81530528a67b448\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c822965e398ee41bc8a5693e14c782abdc26ea1cd24337a9928b56d7632dc3d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af147efcbfe79b90b7cdefa71c882be2ed0bebdbd97936afc5785dbb6dbc99d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://668e9c11154c044fd4eb1970a72735243d021b17e2f048ca2811b3119db0cb00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2rlzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mpwz9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.754367 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p67p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90959adf-c835-40fa-b0f1-26e80426a5c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eafe7f3779b1d7f6511125090218afdeaa7c4a8475e28f3e9bba5db2375f025c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rg5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p67p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.768707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.768731 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.768740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.768769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.768778 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.778485 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c25daf-d20b-4ef8-97be-1ded1fb4239f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8575a719c1412eefc3e8d8dcd4113e27e767c12f94618ce0ead17d975d1a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c01a6edb178570c46ded802aaaebcc7349cc422090121cd9434b9f1de159e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1084356b1af0c03200e5f34c09d6f6053ffddbecd0f8a5baa89d38f982d12404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dbf0d53bb06b82583f5193b7cfae610eb438d3c868c652960333818cfd62ce1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb620ca301f7897befb0e636a2eef1bd1d19516b2df0948bd59e29a9074e27a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1273f1cc3a8fed87d1e828c88cfd28e9a9b981d0e3f8b04599ba6aa84855aa81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://734226a35e0dd059307f90ef0282db64400c61a3bbac9628276513c1f05e6a5a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://732fd8104c82fbdad682c9ace25ba010c2a661c671b93306213e4ac597cb359f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.797181 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5118a70b-3f3c-483d-9e77-bb83969ebcf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53380bd8c248e0b05c99e2aa9b0e0f5c626998d0f390a4a36d95204e8ed904df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc6123fc3270f7a8c6a09ccb5df1c52320d53ed01c96381a35476f282c9c7c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81b7cdb685a74437b33340701174bb374694a1d73563f31857d4b3d644e1156\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69953d3b143c6cb0674eb1eb061210fcc54ef613ee4edff5687471eb3624be01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-23T20:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-23T20:08:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.816202 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.836389 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.856236 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a84a2b95cad7178c11d084bfbb3509ff262d2b21e9e292bee17635a33fe11d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://747f74124dfd42e5681a23bde7e181d0e5bfbad3b92ff596922631b2ef62acd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.871516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.871781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.871971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.872137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.872299 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.874146 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55581d55-59b1-4e06-ac8e-9d7f46b0820b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fabc5f78300809af855a47fd22daf0e7d8fc1dc9115fbbf8f0cf903b5189632\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d0dc6756cad6e7e1c42467f92cfc5d1b50fe7a2948aa91069bd99d0a916d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-23T20:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw82p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-h665k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.895661 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ada6b953-f533-4b33-b07e-7e80604fe4a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-23T20:08:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-23T20:08:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hjmwb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-23T20:09:32Z is after 2025-08-24T17:21:41Z" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.975202 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.975248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.975261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.975277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:32 crc kubenswrapper[4726]: I1123 20:09:32.975292 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:32Z","lastTransitionTime":"2025-11-23T20:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.078061 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.078116 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.078132 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.078154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.078171 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.180531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.180563 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.180574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.180589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.180599 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.284015 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.284065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.284116 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.284138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.284155 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.387173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.387251 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.387268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.387289 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.387334 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.489996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.490069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.490087 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.490114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.490135 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.588222 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:33 crc kubenswrapper[4726]: E1123 20:09:33.588403 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.588696 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:33 crc kubenswrapper[4726]: E1123 20:09:33.588812 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.589052 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:33 crc kubenswrapper[4726]: E1123 20:09:33.589207 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.592712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.592759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.592777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.592799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.592816 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.695613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.695674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.695687 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.695722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.695735 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.797945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.797993 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.798004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.798020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.798032 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.900834 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.900919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.900956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.900974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:33 crc kubenswrapper[4726]: I1123 20:09:33.900985 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:33Z","lastTransitionTime":"2025-11-23T20:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.003934 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.004003 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.004020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.004046 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.004062 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.106637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.106733 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.106763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.106820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.106840 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.209439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.209485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.209531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.209565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.209580 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.311919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.311956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.312000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.312020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.312032 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.414670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.414718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.414730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.414745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.414754 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.518321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.518395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.518419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.518457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.518481 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.588841 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:34 crc kubenswrapper[4726]: E1123 20:09:34.589153 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.621007 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.621045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.621053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.621070 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.621082 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.729477 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.729565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.729593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.729625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.729658 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.833019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.833082 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.833105 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.833134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.833155 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.936361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.936439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.936458 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.936482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:34 crc kubenswrapper[4726]: I1123 20:09:34.936499 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:34Z","lastTransitionTime":"2025-11-23T20:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.039706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.039770 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.039788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.039813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.039858 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.142596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.142678 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.142702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.142727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.142744 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.245855 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.245958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.245983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.246013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.246037 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.309215 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.309391 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309449 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.309415355 +0000 UTC m=+147.458456341 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.309566 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.309614 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309634 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.309663 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309714 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.309691283 +0000 UTC m=+147.458732269 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309788 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309830 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309854 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309901 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309904 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.309848618 +0000 UTC m=+147.458889614 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309951 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.30993612 +0000 UTC m=+147.458977106 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309959 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.309999 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.310018 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.310098 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.310069564 +0000 UTC m=+147.459110560 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.349287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.349343 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.349360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.349384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.349401 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.451574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.451626 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.451646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.451667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.451682 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.554562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.554633 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.554650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.555069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.555099 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.589143 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.589242 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.589303 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.589150 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.589481 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:35 crc kubenswrapper[4726]: E1123 20:09:35.589638 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.657803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.657904 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.657923 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.657949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.657967 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.760850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.760921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.760932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.760948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.760958 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.864377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.864459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.864483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.864520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.864545 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.967069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.967157 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.967182 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.967208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:35 crc kubenswrapper[4726]: I1123 20:09:35.967227 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:35Z","lastTransitionTime":"2025-11-23T20:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.070398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.070494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.070545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.070570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.070587 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.174062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.174125 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.174142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.174168 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.174186 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.277858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.277934 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.277952 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.277976 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.278002 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.381282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.381324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.381335 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.381355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.381366 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.484096 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.484188 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.484213 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.484243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.484268 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.587141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.587196 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.587214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.587240 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.587263 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.588348 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:36 crc kubenswrapper[4726]: E1123 20:09:36.588506 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.690294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.690321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.690331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.690343 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.690351 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.793951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.793998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.794035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.794053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.794064 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.896599 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.896672 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.896686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.896706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.896718 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.999277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.999735 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:36 crc kubenswrapper[4726]: I1123 20:09:36.999763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:36.999791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:36.999809 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:36Z","lastTransitionTime":"2025-11-23T20:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.103258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.103307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.103323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.103347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.103364 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.206164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.206220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.206246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.206280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.206305 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.310047 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.310103 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.310120 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.310142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.310158 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.412493 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.412553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.412572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.412595 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.412612 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.515297 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.515361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.515378 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.515401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.515420 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.588893 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.589102 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:37 crc kubenswrapper[4726]: E1123 20:09:37.589179 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:37 crc kubenswrapper[4726]: E1123 20:09:37.589272 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.589337 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:37 crc kubenswrapper[4726]: E1123 20:09:37.589665 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.618493 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.618538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.618555 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.618577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.618593 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.720981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.721017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.721027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.721038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.721047 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.824660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.824723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.824740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.824765 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.824785 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.929333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.929385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.929403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.929427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:37 crc kubenswrapper[4726]: I1123 20:09:37.929444 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:37Z","lastTransitionTime":"2025-11-23T20:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.032937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.033061 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.033082 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.033108 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.033124 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.136598 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.136715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.136785 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.136812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.136828 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.239711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.239780 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.239797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.239820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.239837 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.342660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.342699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.342709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.342724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.342736 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.445723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.445797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.445811 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.445859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.445893 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.549343 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.549420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.549448 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.549478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.549500 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.588590 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:38 crc kubenswrapper[4726]: E1123 20:09:38.589023 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.652664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.652718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.652744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.652776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.652801 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.756162 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.756216 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.756242 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.756271 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.756294 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.858712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.858788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.858810 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.858904 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.858928 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.962506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.962581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.962607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.962640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:38 crc kubenswrapper[4726]: I1123 20:09:38.962665 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:38Z","lastTransitionTime":"2025-11-23T20:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.066114 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.066178 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.066197 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.066226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.066245 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.169220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.169485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.169658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.169815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.169985 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.273509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.273841 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.274097 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.274298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.274486 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.377462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.377524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.377545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.377569 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.377589 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.480517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.480575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.480586 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.480601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.480612 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.582922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.582985 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.583006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.583032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.583075 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.587955 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:39 crc kubenswrapper[4726]: E1123 20:09:39.588140 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.588026 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:39 crc kubenswrapper[4726]: E1123 20:09:39.588317 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.588196 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:39 crc kubenswrapper[4726]: E1123 20:09:39.588476 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.686498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.686607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.686626 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.686650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.686667 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.789401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.789468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.789480 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.789496 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.789508 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.892030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.892090 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.892110 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.892135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.892153 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.925401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.925465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.925482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.925535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 23 20:09:39 crc kubenswrapper[4726]: I1123 20:09:39.925555 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-23T20:09:39Z","lastTransitionTime":"2025-11-23T20:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.006163 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh"] Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.006757 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.010810 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.010844 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.011295 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.012299 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.088495 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=69.088466582 podStartE2EDuration="1m9.088466582s" podCreationTimestamp="2025-11-23 20:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.060737401 +0000 UTC m=+88.209778437" watchObservedRunningTime="2025-11-23 20:09:40.088466582 +0000 UTC m=+88.237507568" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.112565 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=32.112540408 podStartE2EDuration="32.112540408s" podCreationTimestamp="2025-11-23 20:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.089355748 +0000 UTC m=+88.238396734" watchObservedRunningTime="2025-11-23 20:09:40.112540408 +0000 UTC m=+88.261581404" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.165556 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.165706 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.165745 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.165783 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.165815 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.219715 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-mpwz9" podStartSLOduration=68.219692982 podStartE2EDuration="1m8.219692982s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.204660348 +0000 UTC m=+88.353701364" watchObservedRunningTime="2025-11-23 20:09:40.219692982 +0000 UTC m=+88.368733948" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.239904 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-h665k" podStartSLOduration=67.239849934 podStartE2EDuration="1m7.239849934s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.239502164 +0000 UTC m=+88.388543140" watchObservedRunningTime="2025-11-23 20:09:40.239849934 +0000 UTC m=+88.388890930" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.240228 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-p67p9" podStartSLOduration=68.240219205 podStartE2EDuration="1m8.240219205s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.220346021 +0000 UTC m=+88.369387017" watchObservedRunningTime="2025-11-23 20:09:40.240219205 +0000 UTC m=+88.389260201" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.266301 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.266569 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.266709 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.266808 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.266890 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.267020 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.267087 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.267301 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.276379 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.278765 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=14.278740768 podStartE2EDuration="14.278740768s" podCreationTimestamp="2025-11-23 20:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.277649926 +0000 UTC m=+88.426690912" watchObservedRunningTime="2025-11-23 20:09:40.278740768 +0000 UTC m=+88.427781754" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.297847 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d7e4a47-0976-46a4-9b2d-4da9bb7b1507-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8nblh\" (UID: \"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.299984 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.299964281 podStartE2EDuration="1m9.299964281s" podCreationTimestamp="2025-11-23 20:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.298621112 +0000 UTC m=+88.447662098" watchObservedRunningTime="2025-11-23 20:09:40.299964281 +0000 UTC m=+88.449005277" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.329502 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.350196 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jttr8" podStartSLOduration=68.350176601 podStartE2EDuration="1m8.350176601s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.349017367 +0000 UTC m=+88.498058383" watchObservedRunningTime="2025-11-23 20:09:40.350176601 +0000 UTC m=+88.499217567" Nov 23 20:09:40 crc kubenswrapper[4726]: W1123 20:09:40.362056 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d7e4a47_0976_46a4_9b2d_4da9bb7b1507.slice/crio-5c42977c78c89019ad43bc4ddb17962e3db4d9bdc2f8f863dd378644478a7f99 WatchSource:0}: Error finding container 5c42977c78c89019ad43bc4ddb17962e3db4d9bdc2f8f863dd378644478a7f99: Status 404 returned error can't find the container with id 5c42977c78c89019ad43bc4ddb17962e3db4d9bdc2f8f863dd378644478a7f99 Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.374366 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-bq9lq" podStartSLOduration=69.374346769 podStartE2EDuration="1m9.374346769s" podCreationTimestamp="2025-11-23 20:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.361846548 +0000 UTC m=+88.510887534" watchObservedRunningTime="2025-11-23 20:09:40.374346769 +0000 UTC m=+88.523387735" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.448583 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=66.448562262 podStartE2EDuration="1m6.448562262s" podCreationTimestamp="2025-11-23 20:08:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.447037278 +0000 UTC m=+88.596078244" watchObservedRunningTime="2025-11-23 20:09:40.448562262 +0000 UTC m=+88.597603228" Nov 23 20:09:40 crc kubenswrapper[4726]: I1123 20:09:40.588851 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:40 crc kubenswrapper[4726]: E1123 20:09:40.589325 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:41 crc kubenswrapper[4726]: I1123 20:09:41.155400 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" event={"ID":"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507","Type":"ContainerStarted","Data":"f44cc84bf0371b8506c74a98131082529936e482f7d8264983cc54e34e1bd901"} Nov 23 20:09:41 crc kubenswrapper[4726]: I1123 20:09:41.155639 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" event={"ID":"3d7e4a47-0976-46a4-9b2d-4da9bb7b1507","Type":"ContainerStarted","Data":"5c42977c78c89019ad43bc4ddb17962e3db4d9bdc2f8f863dd378644478a7f99"} Nov 23 20:09:41 crc kubenswrapper[4726]: I1123 20:09:41.168228 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podStartSLOduration=69.168198806 podStartE2EDuration="1m9.168198806s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:40.48170813 +0000 UTC m=+88.630749086" watchObservedRunningTime="2025-11-23 20:09:41.168198806 +0000 UTC m=+89.317239802" Nov 23 20:09:41 crc kubenswrapper[4726]: I1123 20:09:41.169727 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8nblh" podStartSLOduration=69.169706559 podStartE2EDuration="1m9.169706559s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:09:41.16660476 +0000 UTC m=+89.315645726" watchObservedRunningTime="2025-11-23 20:09:41.169706559 +0000 UTC m=+89.318747575" Nov 23 20:09:41 crc kubenswrapper[4726]: I1123 20:09:41.588556 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:41 crc kubenswrapper[4726]: I1123 20:09:41.588617 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:41 crc kubenswrapper[4726]: E1123 20:09:41.588721 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:41 crc kubenswrapper[4726]: E1123 20:09:41.588806 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:41 crc kubenswrapper[4726]: I1123 20:09:41.588952 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:41 crc kubenswrapper[4726]: E1123 20:09:41.589043 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:42 crc kubenswrapper[4726]: I1123 20:09:42.589008 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:42 crc kubenswrapper[4726]: E1123 20:09:42.591177 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:43 crc kubenswrapper[4726]: I1123 20:09:43.589017 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:43 crc kubenswrapper[4726]: I1123 20:09:43.589072 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:43 crc kubenswrapper[4726]: I1123 20:09:43.589072 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:43 crc kubenswrapper[4726]: E1123 20:09:43.589954 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:43 crc kubenswrapper[4726]: E1123 20:09:43.590005 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:43 crc kubenswrapper[4726]: E1123 20:09:43.589791 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:44 crc kubenswrapper[4726]: I1123 20:09:44.588400 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:44 crc kubenswrapper[4726]: E1123 20:09:44.588629 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:44 crc kubenswrapper[4726]: I1123 20:09:44.589520 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:09:44 crc kubenswrapper[4726]: E1123 20:09:44.589791 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:09:45 crc kubenswrapper[4726]: I1123 20:09:45.588131 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:45 crc kubenswrapper[4726]: I1123 20:09:45.588219 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:45 crc kubenswrapper[4726]: I1123 20:09:45.588222 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:45 crc kubenswrapper[4726]: E1123 20:09:45.588389 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:45 crc kubenswrapper[4726]: E1123 20:09:45.589266 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:45 crc kubenswrapper[4726]: E1123 20:09:45.589582 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:46 crc kubenswrapper[4726]: I1123 20:09:46.588366 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:46 crc kubenswrapper[4726]: E1123 20:09:46.589170 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:47 crc kubenswrapper[4726]: I1123 20:09:47.588437 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:47 crc kubenswrapper[4726]: I1123 20:09:47.588530 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:47 crc kubenswrapper[4726]: E1123 20:09:47.588609 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:47 crc kubenswrapper[4726]: I1123 20:09:47.588437 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:47 crc kubenswrapper[4726]: E1123 20:09:47.588710 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:47 crc kubenswrapper[4726]: E1123 20:09:47.588939 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:48 crc kubenswrapper[4726]: I1123 20:09:48.589013 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:48 crc kubenswrapper[4726]: E1123 20:09:48.589175 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:49 crc kubenswrapper[4726]: I1123 20:09:49.588790 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:49 crc kubenswrapper[4726]: I1123 20:09:49.588851 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:49 crc kubenswrapper[4726]: E1123 20:09:49.589079 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:49 crc kubenswrapper[4726]: E1123 20:09:49.589255 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:49 crc kubenswrapper[4726]: I1123 20:09:49.588828 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:49 crc kubenswrapper[4726]: E1123 20:09:49.589699 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:50 crc kubenswrapper[4726]: I1123 20:09:50.589053 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:50 crc kubenswrapper[4726]: E1123 20:09:50.589239 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:50 crc kubenswrapper[4726]: I1123 20:09:50.674037 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:50 crc kubenswrapper[4726]: E1123 20:09:50.674171 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:09:50 crc kubenswrapper[4726]: E1123 20:09:50.674231 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs podName:ada6b953-f533-4b33-b07e-7e80604fe4a1 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:54.674212356 +0000 UTC m=+162.823253322 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs") pod "network-metrics-daemon-hjmwb" (UID: "ada6b953-f533-4b33-b07e-7e80604fe4a1") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 23 20:09:51 crc kubenswrapper[4726]: I1123 20:09:51.588328 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:51 crc kubenswrapper[4726]: I1123 20:09:51.588406 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:51 crc kubenswrapper[4726]: I1123 20:09:51.588406 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:51 crc kubenswrapper[4726]: E1123 20:09:51.589139 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:51 crc kubenswrapper[4726]: E1123 20:09:51.589227 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:51 crc kubenswrapper[4726]: E1123 20:09:51.589846 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:52 crc kubenswrapper[4726]: I1123 20:09:52.589048 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:52 crc kubenswrapper[4726]: E1123 20:09:52.592276 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:53 crc kubenswrapper[4726]: I1123 20:09:53.588637 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:53 crc kubenswrapper[4726]: I1123 20:09:53.588728 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:53 crc kubenswrapper[4726]: E1123 20:09:53.588803 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:53 crc kubenswrapper[4726]: E1123 20:09:53.588977 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:53 crc kubenswrapper[4726]: I1123 20:09:53.589081 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:53 crc kubenswrapper[4726]: E1123 20:09:53.589185 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:54 crc kubenswrapper[4726]: I1123 20:09:54.588763 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:54 crc kubenswrapper[4726]: E1123 20:09:54.588995 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:55 crc kubenswrapper[4726]: I1123 20:09:55.588800 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:55 crc kubenswrapper[4726]: I1123 20:09:55.588850 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:55 crc kubenswrapper[4726]: E1123 20:09:55.589068 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:55 crc kubenswrapper[4726]: I1123 20:09:55.589112 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:55 crc kubenswrapper[4726]: E1123 20:09:55.589291 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:55 crc kubenswrapper[4726]: E1123 20:09:55.589472 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:56 crc kubenswrapper[4726]: I1123 20:09:56.588389 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:56 crc kubenswrapper[4726]: E1123 20:09:56.588586 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:57 crc kubenswrapper[4726]: I1123 20:09:57.588615 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:57 crc kubenswrapper[4726]: I1123 20:09:57.588652 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:57 crc kubenswrapper[4726]: I1123 20:09:57.588852 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:57 crc kubenswrapper[4726]: E1123 20:09:57.589033 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:57 crc kubenswrapper[4726]: E1123 20:09:57.589159 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:57 crc kubenswrapper[4726]: E1123 20:09:57.589741 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:09:57 crc kubenswrapper[4726]: I1123 20:09:57.590266 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:09:57 crc kubenswrapper[4726]: E1123 20:09:57.590578 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-td7h8_openshift-ovn-kubernetes(db0ef36c-6b04-4a1f-bb8d-8f0895fd33be)\"" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" Nov 23 20:09:58 crc kubenswrapper[4726]: I1123 20:09:58.588601 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:09:58 crc kubenswrapper[4726]: E1123 20:09:58.588821 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:09:59 crc kubenswrapper[4726]: I1123 20:09:59.588187 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:09:59 crc kubenswrapper[4726]: I1123 20:09:59.588187 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:09:59 crc kubenswrapper[4726]: E1123 20:09:59.588402 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:09:59 crc kubenswrapper[4726]: I1123 20:09:59.588218 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:09:59 crc kubenswrapper[4726]: E1123 20:09:59.588474 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:09:59 crc kubenswrapper[4726]: E1123 20:09:59.588661 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:00 crc kubenswrapper[4726]: I1123 20:10:00.589346 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:00 crc kubenswrapper[4726]: E1123 20:10:00.589530 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:01 crc kubenswrapper[4726]: I1123 20:10:01.588940 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:01 crc kubenswrapper[4726]: I1123 20:10:01.589009 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:01 crc kubenswrapper[4726]: I1123 20:10:01.589031 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:01 crc kubenswrapper[4726]: E1123 20:10:01.589177 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:01 crc kubenswrapper[4726]: E1123 20:10:01.589313 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:01 crc kubenswrapper[4726]: E1123 20:10:01.589439 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:02 crc kubenswrapper[4726]: I1123 20:10:02.590215 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:02 crc kubenswrapper[4726]: E1123 20:10:02.590453 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:03 crc kubenswrapper[4726]: I1123 20:10:03.588536 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:03 crc kubenswrapper[4726]: I1123 20:10:03.588564 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:03 crc kubenswrapper[4726]: I1123 20:10:03.588586 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:03 crc kubenswrapper[4726]: E1123 20:10:03.588675 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:03 crc kubenswrapper[4726]: E1123 20:10:03.588807 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:03 crc kubenswrapper[4726]: E1123 20:10:03.588939 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:04 crc kubenswrapper[4726]: I1123 20:10:04.588508 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:04 crc kubenswrapper[4726]: E1123 20:10:04.588744 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:05 crc kubenswrapper[4726]: I1123 20:10:05.588782 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:05 crc kubenswrapper[4726]: I1123 20:10:05.588823 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:05 crc kubenswrapper[4726]: E1123 20:10:05.588973 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:05 crc kubenswrapper[4726]: I1123 20:10:05.589059 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:05 crc kubenswrapper[4726]: E1123 20:10:05.589201 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:05 crc kubenswrapper[4726]: E1123 20:10:05.589451 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:06 crc kubenswrapper[4726]: I1123 20:10:06.588827 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:06 crc kubenswrapper[4726]: E1123 20:10:06.588979 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.588157 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.588283 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:07 crc kubenswrapper[4726]: E1123 20:10:07.588567 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.588680 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:07 crc kubenswrapper[4726]: E1123 20:10:07.588746 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:07 crc kubenswrapper[4726]: E1123 20:10:07.588796 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.592458 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/1.log" Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.593234 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/0.log" Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.593305 4726 generic.go:334] "Generic (PLEG): container finished" podID="8c5753d1-1307-4d28-b08f-3588c14ec1bd" containerID="268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3" exitCode=1 Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.593370 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jttr8" event={"ID":"8c5753d1-1307-4d28-b08f-3588c14ec1bd","Type":"ContainerDied","Data":"268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3"} Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.593435 4726 scope.go:117] "RemoveContainer" containerID="0a3f1235d58833becf2fe3b787d57d77b5702ce738e5c59cb7a43b7c62f955ec" Nov 23 20:10:07 crc kubenswrapper[4726]: I1123 20:10:07.593964 4726 scope.go:117] "RemoveContainer" containerID="268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3" Nov 23 20:10:07 crc kubenswrapper[4726]: E1123 20:10:07.594321 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-jttr8_openshift-multus(8c5753d1-1307-4d28-b08f-3588c14ec1bd)\"" pod="openshift-multus/multus-jttr8" podUID="8c5753d1-1307-4d28-b08f-3588c14ec1bd" Nov 23 20:10:08 crc kubenswrapper[4726]: I1123 20:10:08.588927 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:08 crc kubenswrapper[4726]: E1123 20:10:08.589159 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:08 crc kubenswrapper[4726]: I1123 20:10:08.596825 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/1.log" Nov 23 20:10:09 crc kubenswrapper[4726]: I1123 20:10:09.588838 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:09 crc kubenswrapper[4726]: I1123 20:10:09.588951 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:09 crc kubenswrapper[4726]: E1123 20:10:09.588974 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:09 crc kubenswrapper[4726]: I1123 20:10:09.589013 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:09 crc kubenswrapper[4726]: E1123 20:10:09.589106 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:09 crc kubenswrapper[4726]: E1123 20:10:09.589148 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:10 crc kubenswrapper[4726]: I1123 20:10:10.588228 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:10 crc kubenswrapper[4726]: E1123 20:10:10.588963 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:10 crc kubenswrapper[4726]: I1123 20:10:10.589454 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.453457 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hjmwb"] Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.453854 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:11 crc kubenswrapper[4726]: E1123 20:10:11.453970 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.588770 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.588838 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:11 crc kubenswrapper[4726]: E1123 20:10:11.588910 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:11 crc kubenswrapper[4726]: E1123 20:10:11.588971 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.588773 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:11 crc kubenswrapper[4726]: E1123 20:10:11.589045 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.607951 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/3.log" Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.610100 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerStarted","Data":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.610599 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:10:11 crc kubenswrapper[4726]: I1123 20:10:11.639450 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podStartSLOduration=99.639433454 podStartE2EDuration="1m39.639433454s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:11.638819636 +0000 UTC m=+119.787860602" watchObservedRunningTime="2025-11-23 20:10:11.639433454 +0000 UTC m=+119.788474410" Nov 23 20:10:12 crc kubenswrapper[4726]: E1123 20:10:12.553375 4726 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 23 20:10:12 crc kubenswrapper[4726]: E1123 20:10:12.670830 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 20:10:13 crc kubenswrapper[4726]: I1123 20:10:13.588709 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:13 crc kubenswrapper[4726]: I1123 20:10:13.588740 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:13 crc kubenswrapper[4726]: I1123 20:10:13.588717 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:13 crc kubenswrapper[4726]: I1123 20:10:13.588705 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:13 crc kubenswrapper[4726]: E1123 20:10:13.588899 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:13 crc kubenswrapper[4726]: E1123 20:10:13.589104 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:13 crc kubenswrapper[4726]: E1123 20:10:13.589307 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:13 crc kubenswrapper[4726]: E1123 20:10:13.589562 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:15 crc kubenswrapper[4726]: I1123 20:10:15.588973 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:15 crc kubenswrapper[4726]: I1123 20:10:15.589028 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:15 crc kubenswrapper[4726]: I1123 20:10:15.589123 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:15 crc kubenswrapper[4726]: I1123 20:10:15.588988 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:15 crc kubenswrapper[4726]: E1123 20:10:15.589218 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:15 crc kubenswrapper[4726]: E1123 20:10:15.589318 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:15 crc kubenswrapper[4726]: E1123 20:10:15.589517 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:15 crc kubenswrapper[4726]: E1123 20:10:15.589753 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:17 crc kubenswrapper[4726]: I1123 20:10:17.588972 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:17 crc kubenswrapper[4726]: I1123 20:10:17.589017 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:17 crc kubenswrapper[4726]: I1123 20:10:17.589103 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:17 crc kubenswrapper[4726]: E1123 20:10:17.589151 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:17 crc kubenswrapper[4726]: I1123 20:10:17.589203 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:17 crc kubenswrapper[4726]: E1123 20:10:17.589353 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:17 crc kubenswrapper[4726]: E1123 20:10:17.589503 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:17 crc kubenswrapper[4726]: E1123 20:10:17.589660 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:17 crc kubenswrapper[4726]: E1123 20:10:17.672318 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 20:10:19 crc kubenswrapper[4726]: I1123 20:10:19.588860 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:19 crc kubenswrapper[4726]: I1123 20:10:19.588979 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:19 crc kubenswrapper[4726]: E1123 20:10:19.589022 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:19 crc kubenswrapper[4726]: I1123 20:10:19.589086 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:19 crc kubenswrapper[4726]: E1123 20:10:19.589128 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:19 crc kubenswrapper[4726]: E1123 20:10:19.589224 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:19 crc kubenswrapper[4726]: I1123 20:10:19.589329 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:19 crc kubenswrapper[4726]: E1123 20:10:19.589521 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:21 crc kubenswrapper[4726]: I1123 20:10:21.588508 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:21 crc kubenswrapper[4726]: I1123 20:10:21.589474 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:21 crc kubenswrapper[4726]: I1123 20:10:21.589669 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:21 crc kubenswrapper[4726]: E1123 20:10:21.589803 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:21 crc kubenswrapper[4726]: I1123 20:10:21.589696 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:21 crc kubenswrapper[4726]: E1123 20:10:21.589652 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:21 crc kubenswrapper[4726]: E1123 20:10:21.590215 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:21 crc kubenswrapper[4726]: E1123 20:10:21.590350 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:22 crc kubenswrapper[4726]: I1123 20:10:22.590589 4726 scope.go:117] "RemoveContainer" containerID="268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3" Nov 23 20:10:22 crc kubenswrapper[4726]: E1123 20:10:22.673538 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 23 20:10:23 crc kubenswrapper[4726]: I1123 20:10:23.588019 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:23 crc kubenswrapper[4726]: E1123 20:10:23.588149 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:23 crc kubenswrapper[4726]: I1123 20:10:23.588192 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:23 crc kubenswrapper[4726]: I1123 20:10:23.588257 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:23 crc kubenswrapper[4726]: E1123 20:10:23.588364 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:23 crc kubenswrapper[4726]: I1123 20:10:23.588637 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:23 crc kubenswrapper[4726]: E1123 20:10:23.588729 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:23 crc kubenswrapper[4726]: E1123 20:10:23.589014 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:23 crc kubenswrapper[4726]: I1123 20:10:23.658386 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/1.log" Nov 23 20:10:23 crc kubenswrapper[4726]: I1123 20:10:23.658496 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jttr8" event={"ID":"8c5753d1-1307-4d28-b08f-3588c14ec1bd","Type":"ContainerStarted","Data":"57a5a173a024da59a8d4855dfecd282479d0835d0fcbcce8009f17c4cb39b643"} Nov 23 20:10:25 crc kubenswrapper[4726]: I1123 20:10:25.588495 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:25 crc kubenswrapper[4726]: I1123 20:10:25.588561 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:25 crc kubenswrapper[4726]: I1123 20:10:25.588604 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:25 crc kubenswrapper[4726]: E1123 20:10:25.588684 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:25 crc kubenswrapper[4726]: I1123 20:10:25.588735 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:25 crc kubenswrapper[4726]: E1123 20:10:25.588965 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:25 crc kubenswrapper[4726]: E1123 20:10:25.589098 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:25 crc kubenswrapper[4726]: E1123 20:10:25.589193 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:27 crc kubenswrapper[4726]: I1123 20:10:27.588735 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:27 crc kubenswrapper[4726]: I1123 20:10:27.588782 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:27 crc kubenswrapper[4726]: I1123 20:10:27.588789 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:27 crc kubenswrapper[4726]: I1123 20:10:27.588746 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:27 crc kubenswrapper[4726]: E1123 20:10:27.589023 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 23 20:10:27 crc kubenswrapper[4726]: E1123 20:10:27.589175 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hjmwb" podUID="ada6b953-f533-4b33-b07e-7e80604fe4a1" Nov 23 20:10:27 crc kubenswrapper[4726]: E1123 20:10:27.589365 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 23 20:10:27 crc kubenswrapper[4726]: E1123 20:10:27.589438 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.588502 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.588525 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.588564 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.588608 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.593247 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.593398 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.593465 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.593478 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.593525 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 23 20:10:29 crc kubenswrapper[4726]: I1123 20:10:29.593837 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.373522 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.407935 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-m5z9k"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.408519 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.411357 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vfgwr"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.411723 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.412403 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.412540 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.412730 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.414244 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7ngz8"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.414541 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.415018 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.415019 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.415622 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.416058 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.416673 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8phvk"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.417069 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.417663 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.418102 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.419341 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.419695 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.420290 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dvnr5"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.420588 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.423624 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.424504 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.425396 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.426041 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.429644 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.430276 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.431360 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.431748 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.436655 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-llmxt"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.439336 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-llmxt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.439578 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.441669 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.442170 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.448769 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.448878 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.452220 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.452367 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.452449 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.452668 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.454125 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.455508 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.455967 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456029 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456119 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456234 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456290 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456570 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.455976 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456646 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456779 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456797 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456917 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.456930 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457034 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457153 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457225 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457364 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457433 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457470 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457580 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457665 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.457937 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.458007 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.458216 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.458495 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.458528 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.458586 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.458951 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.459024 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.462602 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-q4zwl"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.463041 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vmvvq"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.463339 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.463636 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.463690 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.464930 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465094 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465226 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465319 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465407 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465500 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465624 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465722 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465841 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.465963 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466055 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466141 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466245 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466330 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466426 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466513 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466433 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-m5z9k"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466617 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466795 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466922 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.467049 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.467257 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.467348 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.467439 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.467532 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.467625 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.467710 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.467800 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.469219 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.466761 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.470112 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.471658 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.471695 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.475618 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.478997 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7ngz8"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.485077 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.487054 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.489047 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.512244 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.527755 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vfgwr"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.529185 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vg8gw"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.529819 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.530147 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.530798 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.538495 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.557052 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-q4zwl"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.557183 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z8zkl"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.558401 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.571224 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.571393 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-etcd-client\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.571509 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd8c701f-32dc-479c-a956-b8cbd83def11-audit-dir\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.571629 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.571744 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-config\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.572310 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.572453 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-client-ca\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.572568 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8bfs\" (UniqueName: \"kubernetes.io/projected/7a3756f7-3ccd-4b4b-a888-eb64139a89a3-kube-api-access-g8bfs\") pod \"downloads-7954f5f757-llmxt\" (UID: \"7a3756f7-3ccd-4b4b-a888-eb64139a89a3\") " pod="openshift-console/downloads-7954f5f757-llmxt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.572682 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xpmx\" (UniqueName: \"kubernetes.io/projected/e01392f4-eeca-4d54-89eb-2b45c1a807d4-kube-api-access-9xpmx\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.572794 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcwsh\" (UniqueName: \"kubernetes.io/projected/1d0494d7-1435-4e2b-af86-b30cd6b4f946-kube-api-access-fcwsh\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.572914 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-encryption-config\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573037 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d0494d7-1435-4e2b-af86-b30cd6b4f946-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573158 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e01392f4-eeca-4d54-89eb-2b45c1a807d4-config\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573268 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/002cee03-8433-4e0d-b648-45300fadd459-config\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573370 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-machine-approver-tls\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573480 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573620 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqksm\" (UniqueName: \"kubernetes.io/projected/8f8151ab-3946-4933-8d49-05d7ecb5209c-kube-api-access-pqksm\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573731 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-auth-proxy-config\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573841 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-etcd-client\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.573954 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-policies\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574060 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574150 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k8gh\" (UniqueName: \"kubernetes.io/projected/002cee03-8433-4e0d-b648-45300fadd459-kube-api-access-9k8gh\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574246 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfdrp\" (UniqueName: \"kubernetes.io/projected/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-kube-api-access-lfdrp\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574348 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574450 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-client-ca\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574547 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e01392f4-eeca-4d54-89eb-2b45c1a807d4-serving-cert\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574648 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574743 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574885 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjc9n\" (UniqueName: \"kubernetes.io/projected/b896681a-7b17-4f20-95a2-f0232ef91f3d-kube-api-access-qjc9n\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.574990 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-config\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575101 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc404f4-e217-4063-bb48-4d06d0802f97-serving-cert\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575208 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-config\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575307 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-audit-policies\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575412 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/002cee03-8433-4e0d-b648-45300fadd459-images\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575525 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-service-ca\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575628 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-oauth-serving-cert\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575741 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-serving-cert\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575919 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e01392f4-eeca-4d54-89eb-2b45c1a807d4-trusted-ca\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.575977 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576003 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b896681a-7b17-4f20-95a2-f0232ef91f3d-serving-cert\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576034 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576065 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kbbd\" (UniqueName: \"kubernetes.io/projected/8dc404f4-e217-4063-bb48-4d06d0802f97-kube-api-access-8kbbd\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576096 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-encryption-config\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576120 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576221 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqzqn\" (UniqueName: \"kubernetes.io/projected/187fba44-db16-4325-9df5-75ab2f9da2d3-kube-api-access-fqzqn\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576250 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f8151ab-3946-4933-8d49-05d7ecb5209c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-config\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576332 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/59a6080f-c880-4307-ae04-090a526e528d-audit-dir\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576360 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-config\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576536 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8151ab-3946-4933-8d49-05d7ecb5209c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576772 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f9aa3e3-b807-454b-ae47-783a582350c8-serving-cert\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576807 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td5r5\" (UniqueName: \"kubernetes.io/projected/7445d8ee-ffb1-4610-ae2d-cb511edf1525-kube-api-access-td5r5\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.576830 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8dc404f4-e217-4063-bb48-4d06d0802f97-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.580458 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-serving-cert\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.580639 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.580691 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9bkd\" (UniqueName: \"kubernetes.io/projected/81862660-59a0-461e-b7b1-bf53c6e7a5e3-kube-api-access-c9bkd\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.580802 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-trusted-ca-bundle\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.580841 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-image-import-ca\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81862660-59a0-461e-b7b1-bf53c6e7a5e3-serving-cert\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581098 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-etcd-serving-ca\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581133 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-dir\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581164 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581193 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-oauth-config\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581219 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/59a6080f-c880-4307-ae04-090a526e528d-node-pullsecrets\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581283 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581512 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-service-ca-bundle\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581543 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/83b28544-0d5c-4c19-87cf-ee6fc37c4ed4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w77x8\" (UID: \"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581603 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581710 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbm96\" (UniqueName: \"kubernetes.io/projected/5f9aa3e3-b807-454b-ae47-783a582350c8-kube-api-access-rbm96\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581794 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr4pw\" (UniqueName: \"kubernetes.io/projected/dd8c701f-32dc-479c-a956-b8cbd83def11-kube-api-access-xr4pw\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581823 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/002cee03-8433-4e0d-b648-45300fadd459-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.581848 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpd66\" (UniqueName: \"kubernetes.io/projected/83b28544-0d5c-4c19-87cf-ee6fc37c4ed4-kube-api-access-kpd66\") pod \"cluster-samples-operator-665b6dd947-w77x8\" (UID: \"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.584092 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d0494d7-1435-4e2b-af86-b30cd6b4f946-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.584122 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wbdt\" (UniqueName: \"kubernetes.io/projected/59a6080f-c880-4307-ae04-090a526e528d-kube-api-access-8wbdt\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.584154 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-audit\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.584181 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.584254 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-serving-cert\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.584293 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-config\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.608317 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.611290 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.612143 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.619212 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.619695 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.619722 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.619792 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.619707 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.619707 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.619998 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.620657 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.621203 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.621619 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.621722 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.621816 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.621986 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.622126 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.622242 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.622285 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.622381 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.622666 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.622978 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.636938 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.637444 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.637907 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.638176 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.638529 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.642818 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vg8gw"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.642942 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.655443 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.655915 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.656043 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.655628 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.655636 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.656496 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.656526 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.657586 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.658379 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.658719 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.658736 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.663037 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.664844 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.666678 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.671096 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.674511 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vmvvq"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.675067 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.676619 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.678514 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8phvk"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.678615 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.679084 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jsqsq"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.680027 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.680837 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.683976 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.684618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696154 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8151ab-3946-4933-8d49-05d7ecb5209c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696213 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8dc404f4-e217-4063-bb48-4d06d0802f97-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696243 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f9aa3e3-b807-454b-ae47-783a582350c8-serving-cert\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696265 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td5r5\" (UniqueName: \"kubernetes.io/projected/7445d8ee-ffb1-4610-ae2d-cb511edf1525-kube-api-access-td5r5\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696289 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696314 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-serving-cert\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696337 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9bkd\" (UniqueName: \"kubernetes.io/projected/81862660-59a0-461e-b7b1-bf53c6e7a5e3-kube-api-access-c9bkd\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696355 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-trusted-ca-bundle\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696378 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-image-import-ca\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696418 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81862660-59a0-461e-b7b1-bf53c6e7a5e3-serving-cert\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696434 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-etcd-serving-ca\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696456 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-dir\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696475 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696502 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-oauth-config\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/59a6080f-c880-4307-ae04-090a526e528d-node-pullsecrets\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696546 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696566 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-service-ca-bundle\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696587 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/83b28544-0d5c-4c19-87cf-ee6fc37c4ed4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w77x8\" (UID: \"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696618 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696649 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbm96\" (UniqueName: \"kubernetes.io/projected/5f9aa3e3-b807-454b-ae47-783a582350c8-kube-api-access-rbm96\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696739 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr4pw\" (UniqueName: \"kubernetes.io/projected/dd8c701f-32dc-479c-a956-b8cbd83def11-kube-api-access-xr4pw\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696779 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/002cee03-8433-4e0d-b648-45300fadd459-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696804 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wbdt\" (UniqueName: \"kubernetes.io/projected/59a6080f-c880-4307-ae04-090a526e528d-kube-api-access-8wbdt\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696829 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpd66\" (UniqueName: \"kubernetes.io/projected/83b28544-0d5c-4c19-87cf-ee6fc37c4ed4-kube-api-access-kpd66\") pod \"cluster-samples-operator-665b6dd947-w77x8\" (UID: \"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696853 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d0494d7-1435-4e2b-af86-b30cd6b4f946-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696899 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696926 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-audit\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696947 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-serving-cert\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696964 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-config\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.696985 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697007 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-etcd-client\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697025 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd8c701f-32dc-479c-a956-b8cbd83def11-audit-dir\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697058 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697077 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697117 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-config\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697137 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8bfs\" (UniqueName: \"kubernetes.io/projected/7a3756f7-3ccd-4b4b-a888-eb64139a89a3-kube-api-access-g8bfs\") pod \"downloads-7954f5f757-llmxt\" (UID: \"7a3756f7-3ccd-4b4b-a888-eb64139a89a3\") " pod="openshift-console/downloads-7954f5f757-llmxt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697156 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-client-ca\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697175 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xpmx\" (UniqueName: \"kubernetes.io/projected/e01392f4-eeca-4d54-89eb-2b45c1a807d4-kube-api-access-9xpmx\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-encryption-config\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697219 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmkkd\" (UniqueName: \"kubernetes.io/projected/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-kube-api-access-tmkkd\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697242 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcwsh\" (UniqueName: \"kubernetes.io/projected/1d0494d7-1435-4e2b-af86-b30cd6b4f946-kube-api-access-fcwsh\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697258 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d0494d7-1435-4e2b-af86-b30cd6b4f946-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697276 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e01392f4-eeca-4d54-89eb-2b45c1a807d4-config\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697295 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/002cee03-8433-4e0d-b648-45300fadd459-config\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697313 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-machine-approver-tls\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697333 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697370 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqksm\" (UniqueName: \"kubernetes.io/projected/8f8151ab-3946-4933-8d49-05d7ecb5209c-kube-api-access-pqksm\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697391 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-auth-proxy-config\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697409 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-etcd-client\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697516 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-policies\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697535 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-client-ca\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697699 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k8gh\" (UniqueName: \"kubernetes.io/projected/002cee03-8433-4e0d-b648-45300fadd459-kube-api-access-9k8gh\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697720 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfdrp\" (UniqueName: \"kubernetes.io/projected/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-kube-api-access-lfdrp\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697949 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.697973 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698098 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e01392f4-eeca-4d54-89eb-2b45c1a807d4-serving-cert\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698122 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698267 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-config\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698292 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjc9n\" (UniqueName: \"kubernetes.io/projected/b896681a-7b17-4f20-95a2-f0232ef91f3d-kube-api-access-qjc9n\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698315 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-service-ca\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698394 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc404f4-e217-4063-bb48-4d06d0802f97-serving-cert\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698416 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-config\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698518 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-audit-policies\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698536 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/002cee03-8433-4e0d-b648-45300fadd459-images\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698608 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e01392f4-eeca-4d54-89eb-2b45c1a807d4-trusted-ca\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698629 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698781 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698800 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-oauth-serving-cert\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698953 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-serving-cert\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.698975 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-encryption-config\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.700320 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.701746 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-config\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.702859 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.707664 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8dc404f4-e217-4063-bb48-4d06d0802f97-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.712788 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.712850 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b896681a-7b17-4f20-95a2-f0232ef91f3d-serving-cert\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.712884 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.712913 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.712942 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kbbd\" (UniqueName: \"kubernetes.io/projected/8dc404f4-e217-4063-bb48-4d06d0802f97-kube-api-access-8kbbd\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.712965 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-config\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.712983 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.713002 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqzqn\" (UniqueName: \"kubernetes.io/projected/187fba44-db16-4325-9df5-75ab2f9da2d3-kube-api-access-fqzqn\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.713023 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f8151ab-3946-4933-8d49-05d7ecb5209c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.713071 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/59a6080f-c880-4307-ae04-090a526e528d-audit-dir\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.713089 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-config\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.713677 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.713977 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.718864 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-audit\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.723511 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8151ab-3946-4933-8d49-05d7ecb5209c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.736680 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e01392f4-eeca-4d54-89eb-2b45c1a807d4-trusted-ca\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.738097 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-service-ca\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.738563 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-config\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.738761 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-oauth-serving-cert\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.738811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-audit-policies\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.739139 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-config\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.739368 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.739907 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.739976 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.740245 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f9aa3e3-b807-454b-ae47-783a582350c8-serving-cert\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.740995 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-encryption-config\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.742345 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.743160 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-client-ca\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.743301 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-serving-cert\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.745297 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81862660-59a0-461e-b7b1-bf53c6e7a5e3-serving-cert\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.745757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/83b28544-0d5c-4c19-87cf-ee6fc37c4ed4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w77x8\" (UID: \"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.745778 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/59a6080f-c880-4307-ae04-090a526e528d-audit-dir\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.745789 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.745800 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.746989 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-encryption-config\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.747556 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.747722 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-etcd-serving-ca\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.747742 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/59a6080f-c880-4307-ae04-090a526e528d-node-pullsecrets\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.747851 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-dir\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.748002 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f8151ab-3946-4933-8d49-05d7ecb5209c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.748157 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-config\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.748616 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-image-import-ca\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.748693 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-trusted-ca-bundle\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.749384 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.749533 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd8c701f-32dc-479c-a956-b8cbd83def11-audit-dir\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.749762 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-4wfvf"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.750185 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-service-ca-bundle\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.750260 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.750406 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.750645 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d0494d7-1435-4e2b-af86-b30cd6b4f946-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.751050 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/dd8c701f-32dc-479c-a956-b8cbd83def11-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.751122 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b896681a-7b17-4f20-95a2-f0232ef91f3d-serving-cert\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.751414 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-config\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.751984 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/002cee03-8433-4e0d-b648-45300fadd459-images\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.752259 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-config\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.752392 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-auth-proxy-config\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.752704 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.753339 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-policies\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.754487 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-client-ca\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.754583 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f9aa3e3-b807-454b-ae47-783a582350c8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.755355 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.756182 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-etcd-client\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.756425 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.756711 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/002cee03-8433-4e0d-b648-45300fadd459-config\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.757078 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-oauth-config\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.757333 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d0494d7-1435-4e2b-af86-b30cd6b4f946-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.757718 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.758148 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e01392f4-eeca-4d54-89eb-2b45c1a807d4-config\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.758358 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-machine-approver-tls\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.759967 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59a6080f-c880-4307-ae04-090a526e528d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.760838 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e01392f4-eeca-4d54-89eb-2b45c1a807d4-serving-cert\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.762045 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9q98t"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.762578 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.764208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.764240 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc404f4-e217-4063-bb48-4d06d0802f97-serving-cert\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.764549 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.765664 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mm77f"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.766349 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.768419 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.768920 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.776005 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.777371 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.778399 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.781055 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27hxl"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.782503 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.782757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.783390 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd8c701f-32dc-479c-a956-b8cbd83def11-serving-cert\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.783513 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/002cee03-8433-4e0d-b648-45300fadd459-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.783815 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-serving-cert\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.784461 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/59a6080f-c880-4307-ae04-090a526e528d-etcd-client\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.787093 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.789103 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.789317 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.789823 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.791014 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.791184 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.791584 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.791992 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sn98k"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.792763 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.793137 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dvnr5"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.794574 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-llmxt"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.795624 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.796195 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.797064 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.799346 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z8zkl"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.801049 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.802959 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.804617 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.805821 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.807059 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.808834 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9q98t"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.809982 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jsqsq"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.811137 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.812165 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.813371 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-579gb"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.814154 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-579gb" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.814326 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.814409 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.814549 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.814787 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmkkd\" (UniqueName: \"kubernetes.io/projected/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-kube-api-access-tmkkd\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.814982 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.816513 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hjt8z"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.816636 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.818632 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.825367 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.828483 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.830220 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27hxl"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.830387 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.831682 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mm77f"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.833201 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hjt8z"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.834632 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.836156 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.837546 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-579gb"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.839089 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sn98k"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.840528 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.842016 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.843397 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.846904 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-sbbhd"] Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.847729 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.851109 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.871126 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.890583 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.910164 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.930746 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.951203 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.970734 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 23 20:10:31 crc kubenswrapper[4726]: I1123 20:10:31.990078 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.011837 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.030346 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.051644 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.071381 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.091240 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.111058 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.130592 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.151292 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.171233 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.192361 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.211761 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.231001 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.251247 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.270753 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.291970 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.321242 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-trusted-ca\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.321714 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klphx\" (UniqueName: \"kubernetes.io/projected/0a8210d9-a926-43ad-8830-6b41e594a60e-kube-api-access-klphx\") pod \"dns-operator-744455d44c-z8zkl\" (UID: \"0a8210d9-a926-43ad-8830-6b41e594a60e\") " pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.322094 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.322277 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c928dd1a-1fe0-49d9-8829-608312d1de6f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.322522 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:32.82250027 +0000 UTC m=+140.971541266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.322639 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-certificates\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.322677 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a8210d9-a926-43ad-8830-6b41e594a60e-metrics-tls\") pod \"dns-operator-744455d44c-z8zkl\" (UID: \"0a8210d9-a926-43ad-8830-6b41e594a60e\") " pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.322906 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mxbk\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-kube-api-access-5mxbk\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.322942 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c928dd1a-1fe0-49d9-8829-608312d1de6f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.323098 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-bound-sa-token\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.323151 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-tls\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.331452 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.351648 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.370863 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.391268 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.410776 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424014 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424252 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.424296 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:32.92426307 +0000 UTC m=+141.073304066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424346 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbkbp\" (UniqueName: \"kubernetes.io/projected/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-kube-api-access-nbkbp\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424430 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mxbk\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-kube-api-access-5mxbk\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424462 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-socket-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424488 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c928dd1a-1fe0-49d9-8829-608312d1de6f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlzsh\" (UniqueName: \"kubernetes.io/projected/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-kube-api-access-nlzsh\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424533 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424585 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-metrics-tls\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424606 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3384ef94-8855-487c-9457-1ff0d7a393c9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424706 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5gvl\" (UniqueName: \"kubernetes.io/projected/10063588-a897-402a-b308-7e892b096b33-kube-api-access-s5gvl\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424746 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/581ff47c-292f-4901-8ee4-124258f07696-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424768 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl85r\" (UniqueName: \"kubernetes.io/projected/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-kube-api-access-gl85r\") pod \"ingress-canary-hjt8z\" (UID: \"f62d5e1f-4d77-4474-aa97-6966ad32e2c2\") " pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424798 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-trusted-ca\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36126794-ca3b-42fa-b977-adcc62a6902e-serving-cert\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424845 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10063588-a897-402a-b308-7e892b096b33-proxy-tls\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424888 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424912 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-metrics-certs\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424947 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klphx\" (UniqueName: \"kubernetes.io/projected/0a8210d9-a926-43ad-8830-6b41e594a60e-kube-api-access-klphx\") pod \"dns-operator-744455d44c-z8zkl\" (UID: \"0a8210d9-a926-43ad-8830-6b41e594a60e\") " pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424968 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-ca\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.424992 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-csi-data-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425014 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-node-bootstrap-token\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425034 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkb4w\" (UniqueName: \"kubernetes.io/projected/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-kube-api-access-nkb4w\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425063 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7d578bd-abe7-4012-8f91-04a16dafeb80-tmpfs\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425085 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425107 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-config-volume\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425127 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/561d47d5-3264-46bb-9796-356ab9cfaaf8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27hxl\" (UID: \"561d47d5-3264-46bb-9796-356ab9cfaaf8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425158 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425211 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-plugins-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425232 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425251 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-config\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425271 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425299 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2077c584-5d3d-48bc-9c39-2e5fc941cc38-config\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425327 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-stats-auth\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425327 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c928dd1a-1fe0-49d9-8829-608312d1de6f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425350 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7d578bd-abe7-4012-8f91-04a16dafeb80-apiservice-cert\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425385 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0439b8a1-3e02-42a7-92ba-9a7fca3b3636-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hqxtv\" (UID: \"0439b8a1-3e02-42a7-92ba-9a7fca3b3636\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425416 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-certificates\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.425471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a8210d9-a926-43ad-8830-6b41e594a60e-metrics-tls\") pod \"dns-operator-744455d44c-z8zkl\" (UID: \"0a8210d9-a926-43ad-8830-6b41e594a60e\") " pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.425849 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:32.925832536 +0000 UTC m=+141.074873522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.426390 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-registration-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.426463 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c969e63-8584-4613-85be-1d11f3014d13-secret-volume\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.426674 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-srv-cert\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.426791 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhcv5\" (UniqueName: \"kubernetes.io/projected/36126794-ca3b-42fa-b977-adcc62a6902e-kube-api-access-jhcv5\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.426854 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ea43f27a-45f0-42e3-9c85-360c4e4578cf-signing-key\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.426947 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-certificates\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.426957 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghgvm\" (UniqueName: \"kubernetes.io/projected/178c9142-3c12-40db-a898-81f61fd478f8-kube-api-access-ghgvm\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427011 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427048 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4ff399e8-7c2b-497c-aec8-96383b2f92fb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hlvv6\" (UID: \"4ff399e8-7c2b-497c-aec8-96383b2f92fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427080 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-config\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427112 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/178c9142-3c12-40db-a898-81f61fd478f8-srv-cert\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427233 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psw27\" (UniqueName: \"kubernetes.io/projected/4ff399e8-7c2b-497c-aec8-96383b2f92fb-kube-api-access-psw27\") pod \"control-plane-machine-set-operator-78cbb6b69f-hlvv6\" (UID: \"4ff399e8-7c2b-497c-aec8-96383b2f92fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427278 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ea43f27a-45f0-42e3-9c85-360c4e4578cf-signing-cabundle\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427310 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncpxt\" (UniqueName: \"kubernetes.io/projected/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-kube-api-access-ncpxt\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427580 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-bound-sa-token\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.427828 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-cert\") pod \"ingress-canary-hjt8z\" (UID: \"f62d5e1f-4d77-4474-aa97-6966ad32e2c2\") " pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.428745 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.428787 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-client\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.428829 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-tls\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.428863 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c969e63-8584-4613-85be-1d11f3014d13-config-volume\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.428940 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/581ff47c-292f-4901-8ee4-124258f07696-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-trusted-ca\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429056 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429102 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-certs\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429154 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-846vj\" (UniqueName: \"kubernetes.io/projected/a7d578bd-abe7-4012-8f91-04a16dafeb80-kube-api-access-846vj\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429198 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-images\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429247 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-service-ca\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429285 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwqv2\" (UniqueName: \"kubernetes.io/projected/f8a59726-54f0-4469-b7fc-242607f33e83-kube-api-access-cwqv2\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429336 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg258\" (UniqueName: \"kubernetes.io/projected/0439b8a1-3e02-42a7-92ba-9a7fca3b3636-kube-api-access-pg258\") pod \"package-server-manager-789f6589d5-hqxtv\" (UID: \"0439b8a1-3e02-42a7-92ba-9a7fca3b3636\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429385 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-service-ca-bundle\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429435 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/581ff47c-292f-4901-8ee4-124258f07696-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429541 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/178c9142-3c12-40db-a898-81f61fd478f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429578 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c928dd1a-1fe0-49d9-8829-608312d1de6f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.429601 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hs22\" (UniqueName: \"kubernetes.io/projected/561d47d5-3264-46bb-9796-356ab9cfaaf8-kube-api-access-7hs22\") pod \"multus-admission-controller-857f4d67dd-27hxl\" (UID: \"561d47d5-3264-46bb-9796-356ab9cfaaf8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430055 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7d578bd-abe7-4012-8f91-04a16dafeb80-webhook-cert\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430104 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3384ef94-8855-487c-9457-1ff0d7a393c9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430125 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-proxy-tls\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430145 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3384ef94-8855-487c-9457-1ff0d7a393c9-config\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430164 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-metrics-tls\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430206 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2htmf\" (UniqueName: \"kubernetes.io/projected/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-kube-api-access-2htmf\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430262 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gzwp\" (UniqueName: \"kubernetes.io/projected/2077c584-5d3d-48bc-9c39-2e5fc941cc38-kube-api-access-5gzwp\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430263 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-trusted-ca\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.430284 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2077c584-5d3d-48bc-9c39-2e5fc941cc38-serving-cert\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.431633 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.431664 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk4gx\" (UniqueName: \"kubernetes.io/projected/ea43f27a-45f0-42e3-9c85-360c4e4578cf-kube-api-access-hk4gx\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.431765 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txqc4\" (UniqueName: \"kubernetes.io/projected/8ae6fd55-a974-4701-a37f-b263979a310c-kube-api-access-txqc4\") pod \"migrator-59844c95c7-frgld\" (UID: \"8ae6fd55-a974-4701-a37f-b263979a310c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.431809 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wns9\" (UniqueName: \"kubernetes.io/projected/8c969e63-8584-4613-85be-1d11f3014d13-kube-api-access-5wns9\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.432034 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-default-certificate\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.432076 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhz6c\" (UniqueName: \"kubernetes.io/projected/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-kube-api-access-fhz6c\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.432147 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-mountpoint-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.432201 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6hbr\" (UniqueName: \"kubernetes.io/projected/2030e722-71ba-45d8-8d57-07a317e966e2-kube-api-access-s6hbr\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.432236 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8kzz\" (UniqueName: \"kubernetes.io/projected/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-kube-api-access-k8kzz\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.434241 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-tls\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.436221 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a8210d9-a926-43ad-8830-6b41e594a60e-metrics-tls\") pod \"dns-operator-744455d44c-z8zkl\" (UID: \"0a8210d9-a926-43ad-8830-6b41e594a60e\") " pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.436453 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c928dd1a-1fe0-49d9-8829-608312d1de6f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.451635 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.471264 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.490219 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.533284 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.533517 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.03349206 +0000 UTC m=+141.182533056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.533589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.533630 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbkbp\" (UniqueName: \"kubernetes.io/projected/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-kube-api-access-nbkbp\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.534701 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-socket-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.533675 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-socket-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlzsh\" (UniqueName: \"kubernetes.io/projected/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-kube-api-access-nlzsh\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547183 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547261 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-metrics-tls\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547297 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3384ef94-8855-487c-9457-1ff0d7a393c9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547345 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5gvl\" (UniqueName: \"kubernetes.io/projected/10063588-a897-402a-b308-7e892b096b33-kube-api-access-s5gvl\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547455 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/581ff47c-292f-4901-8ee4-124258f07696-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547498 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl85r\" (UniqueName: \"kubernetes.io/projected/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-kube-api-access-gl85r\") pod \"ingress-canary-hjt8z\" (UID: \"f62d5e1f-4d77-4474-aa97-6966ad32e2c2\") " pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547541 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36126794-ca3b-42fa-b977-adcc62a6902e-serving-cert\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547581 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10063588-a897-402a-b308-7e892b096b33-proxy-tls\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547622 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547667 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-metrics-certs\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547721 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-ca\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547763 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-csi-data-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547796 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-node-bootstrap-token\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547834 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkb4w\" (UniqueName: \"kubernetes.io/projected/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-kube-api-access-nkb4w\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547917 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7d578bd-abe7-4012-8f91-04a16dafeb80-tmpfs\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547962 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.547999 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-config-volume\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548047 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/561d47d5-3264-46bb-9796-356ab9cfaaf8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27hxl\" (UID: \"561d47d5-3264-46bb-9796-356ab9cfaaf8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548102 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548145 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548182 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-plugins-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548227 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548266 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-config\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548307 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548466 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-csi-data-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548487 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2077c584-5d3d-48bc-9c39-2e5fc941cc38-config\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548647 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-stats-auth\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548715 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0439b8a1-3e02-42a7-92ba-9a7fca3b3636-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hqxtv\" (UID: \"0439b8a1-3e02-42a7-92ba-9a7fca3b3636\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548782 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-registration-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548826 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7d578bd-abe7-4012-8f91-04a16dafeb80-apiservice-cert\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548856 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-plugins-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.548907 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c969e63-8584-4613-85be-1d11f3014d13-secret-volume\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549014 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-srv-cert\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549089 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhcv5\" (UniqueName: \"kubernetes.io/projected/36126794-ca3b-42fa-b977-adcc62a6902e-kube-api-access-jhcv5\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549133 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ea43f27a-45f0-42e3-9c85-360c4e4578cf-signing-key\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549185 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghgvm\" (UniqueName: \"kubernetes.io/projected/178c9142-3c12-40db-a898-81f61fd478f8-kube-api-access-ghgvm\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549245 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549291 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4ff399e8-7c2b-497c-aec8-96383b2f92fb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hlvv6\" (UID: \"4ff399e8-7c2b-497c-aec8-96383b2f92fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549322 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-ca\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549336 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-config\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549438 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/178c9142-3c12-40db-a898-81f61fd478f8-srv-cert\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549530 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psw27\" (UniqueName: \"kubernetes.io/projected/4ff399e8-7c2b-497c-aec8-96383b2f92fb-kube-api-access-psw27\") pod \"control-plane-machine-set-operator-78cbb6b69f-hlvv6\" (UID: \"4ff399e8-7c2b-497c-aec8-96383b2f92fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549584 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ea43f27a-45f0-42e3-9c85-360c4e4578cf-signing-cabundle\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549627 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncpxt\" (UniqueName: \"kubernetes.io/projected/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-kube-api-access-ncpxt\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549685 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-cert\") pod \"ingress-canary-hjt8z\" (UID: \"f62d5e1f-4d77-4474-aa97-6966ad32e2c2\") " pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549728 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549841 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-client\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549934 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c969e63-8584-4613-85be-1d11f3014d13-config-volume\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.549983 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/581ff47c-292f-4901-8ee4-124258f07696-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550030 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-trusted-ca\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550070 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550117 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-certs\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550177 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-846vj\" (UniqueName: \"kubernetes.io/projected/a7d578bd-abe7-4012-8f91-04a16dafeb80-kube-api-access-846vj\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550224 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-images\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550279 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-service-ca\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550319 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwqv2\" (UniqueName: \"kubernetes.io/projected/f8a59726-54f0-4469-b7fc-242607f33e83-kube-api-access-cwqv2\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550373 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg258\" (UniqueName: \"kubernetes.io/projected/0439b8a1-3e02-42a7-92ba-9a7fca3b3636-kube-api-access-pg258\") pod \"package-server-manager-789f6589d5-hqxtv\" (UID: \"0439b8a1-3e02-42a7-92ba-9a7fca3b3636\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550426 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-service-ca-bundle\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550467 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/581ff47c-292f-4901-8ee4-124258f07696-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550506 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/178c9142-3c12-40db-a898-81f61fd478f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550547 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hs22\" (UniqueName: \"kubernetes.io/projected/561d47d5-3264-46bb-9796-356ab9cfaaf8-kube-api-access-7hs22\") pod \"multus-admission-controller-857f4d67dd-27hxl\" (UID: \"561d47d5-3264-46bb-9796-356ab9cfaaf8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550620 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7d578bd-abe7-4012-8f91-04a16dafeb80-webhook-cert\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550633 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550666 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3384ef94-8855-487c-9457-1ff0d7a393c9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-proxy-tls\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550748 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3384ef94-8855-487c-9457-1ff0d7a393c9-config\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550786 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-metrics-tls\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550842 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2htmf\" (UniqueName: \"kubernetes.io/projected/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-kube-api-access-2htmf\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550914 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gzwp\" (UniqueName: \"kubernetes.io/projected/2077c584-5d3d-48bc-9c39-2e5fc941cc38-kube-api-access-5gzwp\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550957 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2077c584-5d3d-48bc-9c39-2e5fc941cc38-serving-cert\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.550998 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk4gx\" (UniqueName: \"kubernetes.io/projected/ea43f27a-45f0-42e3-9c85-360c4e4578cf-kube-api-access-hk4gx\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.551042 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txqc4\" (UniqueName: \"kubernetes.io/projected/8ae6fd55-a974-4701-a37f-b263979a310c-kube-api-access-txqc4\") pod \"migrator-59844c95c7-frgld\" (UID: \"8ae6fd55-a974-4701-a37f-b263979a310c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.551081 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wns9\" (UniqueName: \"kubernetes.io/projected/8c969e63-8584-4613-85be-1d11f3014d13-kube-api-access-5wns9\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.551124 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhz6c\" (UniqueName: \"kubernetes.io/projected/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-kube-api-access-fhz6c\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.551175 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-mountpoint-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.551215 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6hbr\" (UniqueName: \"kubernetes.io/projected/2030e722-71ba-45d8-8d57-07a317e966e2-kube-api-access-s6hbr\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.551255 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8kzz\" (UniqueName: \"kubernetes.io/projected/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-kube-api-access-k8kzz\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.551296 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-default-certificate\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.553937 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.0539012 +0000 UTC m=+141.202942186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.556273 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-mountpoint-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.557546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2030e722-71ba-45d8-8d57-07a317e966e2-registration-dir\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.557801 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wbdt\" (UniqueName: \"kubernetes.io/projected/59a6080f-c880-4307-ae04-090a526e528d-kube-api-access-8wbdt\") pod \"apiserver-76f77b778f-m5z9k\" (UID: \"59a6080f-c880-4307-ae04-090a526e528d\") " pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.558048 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-trusted-ca\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.558686 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-config\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.559675 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/581ff47c-292f-4901-8ee4-124258f07696-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.559729 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7d578bd-abe7-4012-8f91-04a16dafeb80-tmpfs\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.560421 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3384ef94-8855-487c-9457-1ff0d7a393c9-config\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.561147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.561471 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-config\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.562056 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-service-ca\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.562364 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-metrics-tls\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.563024 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/581ff47c-292f-4901-8ee4-124258f07696-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.563163 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36126794-ca3b-42fa-b977-adcc62a6902e-serving-cert\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.563657 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.563757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c969e63-8584-4613-85be-1d11f3014d13-secret-volume\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.564765 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/36126794-ca3b-42fa-b977-adcc62a6902e-etcd-client\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.566010 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3384ef94-8855-487c-9457-1ff0d7a393c9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.570062 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjc9n\" (UniqueName: \"kubernetes.io/projected/b896681a-7b17-4f20-95a2-f0232ef91f3d-kube-api-access-qjc9n\") pod \"controller-manager-879f6c89f-7ngz8\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.570141 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7d578bd-abe7-4012-8f91-04a16dafeb80-webhook-cert\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.570549 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4ff399e8-7c2b-497c-aec8-96383b2f92fb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hlvv6\" (UID: \"4ff399e8-7c2b-497c-aec8-96383b2f92fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.570672 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-proxy-tls\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.572309 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7d578bd-abe7-4012-8f91-04a16dafeb80-apiservice-cert\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.574262 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-srv-cert\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.576362 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.577741 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/178c9142-3c12-40db-a898-81f61fd478f8-profile-collector-cert\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.581667 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpd66\" (UniqueName: \"kubernetes.io/projected/83b28544-0d5c-4c19-87cf-ee6fc37c4ed4-kube-api-access-kpd66\") pod \"cluster-samples-operator-665b6dd947-w77x8\" (UID: \"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.597273 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbm96\" (UniqueName: \"kubernetes.io/projected/5f9aa3e3-b807-454b-ae47-783a582350c8-kube-api-access-rbm96\") pod \"authentication-operator-69f744f599-8phvk\" (UID: \"5f9aa3e3-b807-454b-ae47-783a582350c8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.614494 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td5r5\" (UniqueName: \"kubernetes.io/projected/7445d8ee-ffb1-4610-ae2d-cb511edf1525-kube-api-access-td5r5\") pod \"console-f9d7485db-q4zwl\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.636155 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kbbd\" (UniqueName: \"kubernetes.io/projected/8dc404f4-e217-4063-bb48-4d06d0802f97-kube-api-access-8kbbd\") pod \"openshift-config-operator-7777fb866f-f8fmr\" (UID: \"8dc404f4-e217-4063-bb48-4d06d0802f97\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.639245 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.650633 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xpmx\" (UniqueName: \"kubernetes.io/projected/e01392f4-eeca-4d54-89eb-2b45c1a807d4-kube-api-access-9xpmx\") pod \"console-operator-58897d9998-dvnr5\" (UID: \"e01392f4-eeca-4d54-89eb-2b45c1a807d4\") " pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.654023 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.654156 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.154132535 +0000 UTC m=+141.303173521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.654813 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.655510 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.155491185 +0000 UTC m=+141.304532171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.672154 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8bfs\" (UniqueName: \"kubernetes.io/projected/7a3756f7-3ccd-4b4b-a888-eb64139a89a3-kube-api-access-g8bfs\") pod \"downloads-7954f5f757-llmxt\" (UID: \"7a3756f7-3ccd-4b4b-a888-eb64139a89a3\") " pod="openshift-console/downloads-7954f5f757-llmxt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.680398 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.699640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqzqn\" (UniqueName: \"kubernetes.io/projected/187fba44-db16-4325-9df5-75ab2f9da2d3-kube-api-access-fqzqn\") pod \"oauth-openshift-558db77b4-vmvvq\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.704794 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.717643 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.730937 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcwsh\" (UniqueName: \"kubernetes.io/projected/1d0494d7-1435-4e2b-af86-b30cd6b4f946-kube-api-access-fcwsh\") pod \"openshift-controller-manager-operator-756b6f6bc6-vfcn6\" (UID: \"1d0494d7-1435-4e2b-af86-b30cd6b4f946\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.734417 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.735089 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9bkd\" (UniqueName: \"kubernetes.io/projected/81862660-59a0-461e-b7b1-bf53c6e7a5e3-kube-api-access-c9bkd\") pod \"route-controller-manager-6576b87f9c-dnsz8\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.739604 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.744467 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-service-ca-bundle\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.752086 4726 request.go:700] Waited for 1.00155756s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-dockercfg-zdk86&limit=500&resourceVersion=0 Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.754271 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.759711 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.760605 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.260583943 +0000 UTC m=+141.409624909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.773028 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.777274 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.779564 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.791149 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.795724 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-default-certificate\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.798488 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.811147 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-llmxt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.813078 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.818469 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-stats-auth\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.824036 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.827551 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-metrics-certs\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.830786 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.852292 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.863222 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.865498 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.365482415 +0000 UTC m=+141.514523371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.869759 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.910271 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr4pw\" (UniqueName: \"kubernetes.io/projected/dd8c701f-32dc-479c-a956-b8cbd83def11-kube-api-access-xr4pw\") pod \"apiserver-7bbb656c7d-z4rlj\" (UID: \"dd8c701f-32dc-479c-a956-b8cbd83def11\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.915077 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqksm\" (UniqueName: \"kubernetes.io/projected/8f8151ab-3946-4933-8d49-05d7ecb5209c-kube-api-access-pqksm\") pod \"openshift-apiserver-operator-796bbdcf4f-t26mb\" (UID: \"8f8151ab-3946-4933-8d49-05d7ecb5209c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.949408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k8gh\" (UniqueName: \"kubernetes.io/projected/002cee03-8433-4e0d-b648-45300fadd459-kube-api-access-9k8gh\") pod \"machine-api-operator-5694c8668f-vfgwr\" (UID: \"002cee03-8433-4e0d-b648-45300fadd459\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.954894 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.956247 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfdrp\" (UniqueName: \"kubernetes.io/projected/c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8-kube-api-access-lfdrp\") pod \"machine-approver-56656f9798-9ch2n\" (UID: \"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.957304 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.964935 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.965297 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.465270307 +0000 UTC m=+141.614311263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.965472 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:32 crc kubenswrapper[4726]: E1123 20:10:32.966211 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.466199325 +0000 UTC m=+141.615240281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.974026 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.980577 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 23 20:10:32 crc kubenswrapper[4726]: I1123 20:10:32.994848 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.002977 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.007899 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ea43f27a-45f0-42e3-9c85-360c4e4578cf-signing-key\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.012029 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.014175 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ea43f27a-45f0-42e3-9c85-360c4e4578cf-signing-cabundle\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.023124 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.031305 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8phvk"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.031466 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: W1123 20:10:33.034019 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc50d7dd0_c9f5_4dc6_87df_27c70cf80ed8.slice/crio-e3e25f92ce187bf461272a7df5bbca9b9e793ae365af0d63dcf57c7fe5ae2aa8 WatchSource:0}: Error finding container e3e25f92ce187bf461272a7df5bbca9b9e793ae365af0d63dcf57c7fe5ae2aa8: Status 404 returned error can't find the container with id e3e25f92ce187bf461272a7df5bbca9b9e793ae365af0d63dcf57c7fe5ae2aa8 Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.052314 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.067789 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.068384 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.568369657 +0000 UTC m=+141.717410613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.072699 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 23 20:10:33 crc kubenswrapper[4726]: W1123 20:10:33.089186 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f9aa3e3_b807_454b_ae47_783a582350c8.slice/crio-bc2114579da950556a62d439c85494a4ce084635e63cb6928b8e58961662fcd5 WatchSource:0}: Error finding container bc2114579da950556a62d439c85494a4ce084635e63cb6928b8e58961662fcd5: Status 404 returned error can't find the container with id bc2114579da950556a62d439c85494a4ce084635e63cb6928b8e58961662fcd5 Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.091940 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.100017 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.121752 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.135383 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.142936 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-m5z9k"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.144186 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.151317 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7ngz8"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.152910 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.161455 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0439b8a1-3e02-42a7-92ba-9a7fca3b3636-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hqxtv\" (UID: \"0439b8a1-3e02-42a7-92ba-9a7fca3b3636\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.169888 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.170381 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.670369454 +0000 UTC m=+141.819410410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.172910 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.178248 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.192291 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.212619 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.239646 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.249025 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.250407 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.270583 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.271446 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.771432584 +0000 UTC m=+141.920473540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.272774 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.284322 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/561d47d5-3264-46bb-9796-356ab9cfaaf8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27hxl\" (UID: \"561d47d5-3264-46bb-9796-356ab9cfaaf8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.290882 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.307461 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-q4zwl"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.312491 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.316408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/178c9142-3c12-40db-a898-81f61fd478f8-srv-cert\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.336678 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.339965 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dvnr5"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.341187 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.345934 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c969e63-8584-4613-85be-1d11f3014d13-config-volume\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.354743 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.356737 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-llmxt"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.357911 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr"] Nov 23 20:10:33 crc kubenswrapper[4726]: W1123 20:10:33.368001 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a3756f7_3ccd_4b4b_a888_eb64139a89a3.slice/crio-b1a31046becdb644d9bc6c03620045e4cefdb6ee28f618ecc78a47dfa1d66f9d WatchSource:0}: Error finding container b1a31046becdb644d9bc6c03620045e4cefdb6ee28f618ecc78a47dfa1d66f9d: Status 404 returned error can't find the container with id b1a31046becdb644d9bc6c03620045e4cefdb6ee28f618ecc78a47dfa1d66f9d Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.372285 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.373513 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.375335 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.875313676 +0000 UTC m=+142.024354632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.394437 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.396656 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.410165 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 23 20:10:33 crc kubenswrapper[4726]: W1123 20:10:33.418267 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f8151ab_3946_4933_8d49_05d7ecb5209c.slice/crio-7341a0f9c8afa69aafaf7410e76ecb2e6217a851a8b102607e591d4372f54a20 WatchSource:0}: Error finding container 7341a0f9c8afa69aafaf7410e76ecb2e6217a851a8b102607e591d4372f54a20: Status 404 returned error can't find the container with id 7341a0f9c8afa69aafaf7410e76ecb2e6217a851a8b102607e591d4372f54a20 Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.431004 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.435377 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2077c584-5d3d-48bc-9c39-2e5fc941cc38-serving-cert\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.439409 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2077c584-5d3d-48bc-9c39-2e5fc941cc38-config\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.453568 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.453791 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vmvvq"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.460062 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.460099 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.471401 4726 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.474417 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.475244 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:33.975228182 +0000 UTC m=+142.124269138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: W1123 20:10:33.477303 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d0494d7_1435_4e2b_af86_b30cd6b4f946.slice/crio-0adbc79db5c78bb1403dec0abb18fd5e78d7f727dea1ad7ec3f38c22600adcb4 WatchSource:0}: Error finding container 0adbc79db5c78bb1403dec0abb18fd5e78d7f727dea1ad7ec3f38c22600adcb4: Status 404 returned error can't find the container with id 0adbc79db5c78bb1403dec0abb18fd5e78d7f727dea1ad7ec3f38c22600adcb4 Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.490338 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.509860 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.530190 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.549859 4726 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.549951 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-node-bootstrap-token podName:0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.049931388 +0000 UTC m=+142.198972344 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-node-bootstrap-token") pod "machine-config-server-sbbhd" (UID: "0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a") : failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.550705 4726 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.550740 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10063588-a897-402a-b308-7e892b096b33-proxy-tls podName:10063588-a897-402a-b308-7e892b096b33 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.050732311 +0000 UTC m=+142.199773267 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/10063588-a897-402a-b308-7e892b096b33-proxy-tls") pod "machine-config-operator-74547568cd-qs2w2" (UID: "10063588-a897-402a-b308-7e892b096b33") : failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.550779 4726 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.550798 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-config-volume podName:b66e5a55-9c4e-4383-9d4f-6961fbd5ab86 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.050791783 +0000 UTC m=+142.199832739 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-config-volume") pod "dns-default-579gb" (UID: "b66e5a55-9c4e-4383-9d4f-6961fbd5ab86") : failed to sync configmap cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.552096 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.554524 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vfgwr"] Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.554568 4726 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.554782 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-cert podName:f62d5e1f-4d77-4474-aa97-6966ad32e2c2 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.0547706 +0000 UTC m=+142.203811556 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-cert") pod "ingress-canary-hjt8z" (UID: "f62d5e1f-4d77-4474-aa97-6966ad32e2c2") : failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.555123 4726 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.555178 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-metrics-tls podName:b66e5a55-9c4e-4383-9d4f-6961fbd5ab86 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.055162511 +0000 UTC m=+142.204203467 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-metrics-tls") pod "dns-default-579gb" (UID: "b66e5a55-9c4e-4383-9d4f-6961fbd5ab86") : failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.557773 4726 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.557813 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-certs podName:0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.057802839 +0000 UTC m=+142.206843805 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-certs") pod "machine-config-server-sbbhd" (UID: "0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a") : failed to sync secret cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.562222 4726 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.562280 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-images podName:10063588-a897-402a-b308-7e892b096b33 nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.06226764 +0000 UTC m=+142.211308596 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-images") pod "machine-config-operator-74547568cd-qs2w2" (UID: "10063588-a897-402a-b308-7e892b096b33") : failed to sync configmap cache: timed out waiting for the condition Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.564454 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj"] Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.570256 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.576160 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.576565 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.07655161 +0000 UTC m=+142.225592566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.610325 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.637217 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.650796 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.679934 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.681188 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.181172634 +0000 UTC m=+142.330213580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.686996 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmkkd\" (UniqueName: \"kubernetes.io/projected/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-kube-api-access-tmkkd\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.716336 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.729499 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-g5k42\" (UID: \"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.734070 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.748341 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" event={"ID":"187fba44-db16-4325-9df5-75ab2f9da2d3","Type":"ContainerStarted","Data":"8a50c562fd384968d05d5fcf20f92b1eb5b5dbcf3eb1866a2ab70537912c0b24"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.757271 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" event={"ID":"1d0494d7-1435-4e2b-af86-b30cd6b4f946","Type":"ContainerStarted","Data":"83267f456749588c552de39e86127d75aa0fdd85e869c1595470ac214397c9c2"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.757317 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" event={"ID":"1d0494d7-1435-4e2b-af86-b30cd6b4f946","Type":"ContainerStarted","Data":"0adbc79db5c78bb1403dec0abb18fd5e78d7f727dea1ad7ec3f38c22600adcb4"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.759394 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-q4zwl" event={"ID":"7445d8ee-ffb1-4610-ae2d-cb511edf1525","Type":"ContainerStarted","Data":"6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.759420 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-q4zwl" event={"ID":"7445d8ee-ffb1-4610-ae2d-cb511edf1525","Type":"ContainerStarted","Data":"8c440bd11882b91bdbf87e47af3c4de4ab04e2d0138649e1caa32f7d2922d045"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.762729 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.765106 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" event={"ID":"8dc404f4-e217-4063-bb48-4d06d0802f97","Type":"ContainerStarted","Data":"b5b3468be68ba899a797074b55a1379732bab40ad5d9883c695ebded00cfe2c2"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.765144 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" event={"ID":"8dc404f4-e217-4063-bb48-4d06d0802f97","Type":"ContainerStarted","Data":"b757514b7f5d3ab08a83cefc4f14401bdb401d69947df68452cf360ee745fe88"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.766733 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" event={"ID":"81862660-59a0-461e-b7b1-bf53c6e7a5e3","Type":"ContainerStarted","Data":"8b55c8593ad2d04c18945d6c7ab850c97c49918e406bafde1f235f0d670dba85"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.766853 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" event={"ID":"81862660-59a0-461e-b7b1-bf53c6e7a5e3","Type":"ContainerStarted","Data":"040440aaa2e63c884c09d474b4fc95644b236c46f494608334925aa2f49539e1"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.767536 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.768445 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" event={"ID":"dd8c701f-32dc-479c-a956-b8cbd83def11","Type":"ContainerStarted","Data":"7d5fa02ff9df1f3c12e0d2160e3affcf7478e570701a1f681329013cda73681a"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.768911 4726 request.go:700] Waited for 1.949752833s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.769731 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" event={"ID":"002cee03-8433-4e0d-b648-45300fadd459","Type":"ContainerStarted","Data":"4c5304ed1fd3a71f61b6a0b1e4b5e728819c851c175f86939aef753e4d9e13e5"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.771034 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.772327 4726 generic.go:334] "Generic (PLEG): container finished" podID="59a6080f-c880-4307-ae04-090a526e528d" containerID="330a476516afe13dfbc9c6e43ea6ca66d4c46f70c5868573e432cff06d72c7fa" exitCode=0 Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.772491 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" event={"ID":"59a6080f-c880-4307-ae04-090a526e528d","Type":"ContainerDied","Data":"330a476516afe13dfbc9c6e43ea6ca66d4c46f70c5868573e432cff06d72c7fa"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.772541 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" event={"ID":"59a6080f-c880-4307-ae04-090a526e528d","Type":"ContainerStarted","Data":"3a3ca7478843ff2dcf7e9937956e1cc211d1c52982a529a74ddded83f22a6e28"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.774893 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-llmxt" event={"ID":"7a3756f7-3ccd-4b4b-a888-eb64139a89a3","Type":"ContainerStarted","Data":"3bbd302901234f48ab9d4d1799fd767957ca548d6529428f73909809eafd4652"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.774920 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-llmxt" event={"ID":"7a3756f7-3ccd-4b4b-a888-eb64139a89a3","Type":"ContainerStarted","Data":"b1a31046becdb644d9bc6c03620045e4cefdb6ee28f618ecc78a47dfa1d66f9d"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.776240 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-llmxt" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.776341 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-llmxt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.776373 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-llmxt" podUID="7a3756f7-3ccd-4b4b-a888-eb64139a89a3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.778754 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" event={"ID":"b896681a-7b17-4f20-95a2-f0232ef91f3d","Type":"ContainerStarted","Data":"446d9a54daf79eb2bbfe87a732ee034e94fbc9ba242853b355bca06640fc9111"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.778836 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" event={"ID":"b896681a-7b17-4f20-95a2-f0232ef91f3d","Type":"ContainerStarted","Data":"e9e1b21e538879a096bd3704701f09c0cb38dd8820f77b9e71af49d413c0e6d6"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.780058 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.781122 4726 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dnsz8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.781150 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" podUID="81862660-59a0-461e-b7b1-bf53c6e7a5e3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.781213 4726 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7ngz8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.781284 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" podUID="b896681a-7b17-4f20-95a2-f0232ef91f3d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.782064 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.782640 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.282629185 +0000 UTC m=+142.431670131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.783844 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" event={"ID":"8f8151ab-3946-4933-8d49-05d7ecb5209c","Type":"ContainerStarted","Data":"1e1a959cad0fb3623e88b5a6260ae7037df6348da212025f9e5d371ab318c225"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.783911 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" event={"ID":"8f8151ab-3946-4933-8d49-05d7ecb5209c","Type":"ContainerStarted","Data":"7341a0f9c8afa69aafaf7410e76ecb2e6217a851a8b102607e591d4372f54a20"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.788142 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" event={"ID":"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8","Type":"ContainerStarted","Data":"236d47095e72e3f2b8a2d478e8a938f00d6a7efa1e40966392de5e30521792e0"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.788182 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" event={"ID":"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8","Type":"ContainerStarted","Data":"e3e25f92ce187bf461272a7df5bbca9b9e793ae365af0d63dcf57c7fe5ae2aa8"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.789992 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.792436 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" event={"ID":"5f9aa3e3-b807-454b-ae47-783a582350c8","Type":"ContainerStarted","Data":"9bf77ff369e08f0491f4ebe081a373399f15823fcaa462f2a135aafb79cf8591"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.792557 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" event={"ID":"5f9aa3e3-b807-454b-ae47-783a582350c8","Type":"ContainerStarted","Data":"bc2114579da950556a62d439c85494a4ce084635e63cb6928b8e58961662fcd5"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.794322 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" event={"ID":"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4","Type":"ContainerStarted","Data":"64bf445be2f15c0df85d7289a93d9f38e45fd9e7728900eaa4ddbcf994f0cb30"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.797601 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" event={"ID":"e01392f4-eeca-4d54-89eb-2b45c1a807d4","Type":"ContainerStarted","Data":"08967e8d2da6cf0f50b576f2ee9167a76d91dcc4701d3bb11ad772b33d76d4c4"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.797636 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" event={"ID":"e01392f4-eeca-4d54-89eb-2b45c1a807d4","Type":"ContainerStarted","Data":"5ee827d06fc017cbb78b891c8bff160788e27376baba3a51f35bb2d7aa02bdbf"} Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.798364 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.801978 4726 patch_prober.go:28] interesting pod/console-operator-58897d9998-dvnr5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.802020 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" podUID="e01392f4-eeca-4d54-89eb-2b45c1a807d4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.810458 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.816019 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.831327 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.864180 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mxbk\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-kube-api-access-5mxbk\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.883001 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.883171 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.383145279 +0000 UTC m=+142.532186235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.883856 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.889642 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.389621949 +0000 UTC m=+142.538662905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.906439 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klphx\" (UniqueName: \"kubernetes.io/projected/0a8210d9-a926-43ad-8830-6b41e594a60e-kube-api-access-klphx\") pod \"dns-operator-744455d44c-z8zkl\" (UID: \"0a8210d9-a926-43ad-8830-6b41e594a60e\") " pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.929018 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mcmsj\" (UID: \"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.929551 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-bound-sa-token\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.943352 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbkbp\" (UniqueName: \"kubernetes.io/projected/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-kube-api-access-nbkbp\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.969349 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlzsh\" (UniqueName: \"kubernetes.io/projected/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-kube-api-access-nlzsh\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.983336 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3384ef94-8855-487c-9457-1ff0d7a393c9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dvndr\" (UID: \"3384ef94-8855-487c-9457-1ff0d7a393c9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.985320 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.985488 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.485463255 +0000 UTC m=+142.634504211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:33 crc kubenswrapper[4726]: I1123 20:10:33.985691 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:33 crc kubenswrapper[4726]: E1123 20:10:33.986245 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.486230798 +0000 UTC m=+142.635271764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.009373 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/581ff47c-292f-4901-8ee4-124258f07696-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cv7p8\" (UID: \"581ff47c-292f-4901-8ee4-124258f07696\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.024715 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5gvl\" (UniqueName: \"kubernetes.io/projected/10063588-a897-402a-b308-7e892b096b33-kube-api-access-s5gvl\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.051476 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl85r\" (UniqueName: \"kubernetes.io/projected/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-kube-api-access-gl85r\") pod \"ingress-canary-hjt8z\" (UID: \"f62d5e1f-4d77-4474-aa97-6966ad32e2c2\") " pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.072563 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkb4w\" (UniqueName: \"kubernetes.io/projected/a9efda09-ccd1-47d3-bdc1-b397b2e751f1-kube-api-access-nkb4w\") pod \"router-default-5444994796-4wfvf\" (UID: \"a9efda09-ccd1-47d3-bdc1-b397b2e751f1\") " pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.083735 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psw27\" (UniqueName: \"kubernetes.io/projected/4ff399e8-7c2b-497c-aec8-96383b2f92fb-kube-api-access-psw27\") pod \"control-plane-machine-set-operator-78cbb6b69f-hlvv6\" (UID: \"4ff399e8-7c2b-497c-aec8-96383b2f92fb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.087664 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.087962 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10063588-a897-402a-b308-7e892b096b33-proxy-tls\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.087997 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-node-bootstrap-token\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.088023 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-config-volume\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.088150 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-cert\") pod \"ingress-canary-hjt8z\" (UID: \"f62d5e1f-4d77-4474-aa97-6966ad32e2c2\") " pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.088190 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-certs\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.088223 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-images\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.088275 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-metrics-tls\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.089208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-config-volume\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.089447 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.589406 +0000 UTC m=+142.738446956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.090509 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42"] Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.091449 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-metrics-tls\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.091592 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/10063588-a897-402a-b308-7e892b096b33-images\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.091709 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10063588-a897-402a-b308-7e892b096b33-proxy-tls\") pod \"machine-config-operator-74547568cd-qs2w2\" (UID: \"10063588-a897-402a-b308-7e892b096b33\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.094177 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-node-bootstrap-token\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.094350 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a-certs\") pod \"machine-config-server-sbbhd\" (UID: \"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a\") " pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.095000 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f62d5e1f-4d77-4474-aa97-6966ad32e2c2-cert\") pod \"ingress-canary-hjt8z\" (UID: \"f62d5e1f-4d77-4474-aa97-6966ad32e2c2\") " pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.111308 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b4e6d66f-e455-4f27-9f46-a03b2f20cb51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hxw6z\" (UID: \"b4e6d66f-e455-4f27-9f46-a03b2f20cb51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.122771 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncpxt\" (UniqueName: \"kubernetes.io/projected/8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa-kube-api-access-ncpxt\") pod \"machine-config-controller-84d6567774-m2fhd\" (UID: \"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.145678 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghgvm\" (UniqueName: \"kubernetes.io/projected/178c9142-3c12-40db-a898-81f61fd478f8-kube-api-access-ghgvm\") pod \"catalog-operator-68c6474976-6jlhs\" (UID: \"178c9142-3c12-40db-a898-81f61fd478f8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.165964 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhcv5\" (UniqueName: \"kubernetes.io/projected/36126794-ca3b-42fa-b977-adcc62a6902e-kube-api-access-jhcv5\") pod \"etcd-operator-b45778765-jsqsq\" (UID: \"36126794-ca3b-42fa-b977-adcc62a6902e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.170962 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.179404 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.185797 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2htmf\" (UniqueName: \"kubernetes.io/projected/8e41e7f4-8c3f-431c-aa5a-fa79132ceef1-kube-api-access-2htmf\") pod \"olm-operator-6b444d44fb-gwlrk\" (UID: \"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.186524 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.190447 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.193275 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.693232981 +0000 UTC m=+142.842273937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.193539 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.200923 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.203853 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gzwp\" (UniqueName: \"kubernetes.io/projected/2077c584-5d3d-48bc-9c39-2e5fc941cc38-kube-api-access-5gzwp\") pod \"service-ca-operator-777779d784-ks9vb\" (UID: \"2077c584-5d3d-48bc-9c39-2e5fc941cc38\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.213196 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.223619 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk4gx\" (UniqueName: \"kubernetes.io/projected/ea43f27a-45f0-42e3-9c85-360c4e4578cf-kube-api-access-hk4gx\") pod \"service-ca-9c57cc56f-9q98t\" (UID: \"ea43f27a-45f0-42e3-9c85-360c4e4578cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.227201 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.242326 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.251593 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.262045 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.264912 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.269405 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txqc4\" (UniqueName: \"kubernetes.io/projected/8ae6fd55-a974-4701-a37f-b263979a310c-kube-api-access-txqc4\") pod \"migrator-59844c95c7-frgld\" (UID: \"8ae6fd55-a974-4701-a37f-b263979a310c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.275447 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wns9\" (UniqueName: \"kubernetes.io/projected/8c969e63-8584-4613-85be-1d11f3014d13-kube-api-access-5wns9\") pod \"collect-profiles-29398800-x676h\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.290054 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhz6c\" (UniqueName: \"kubernetes.io/projected/b66e5a55-9c4e-4383-9d4f-6961fbd5ab86-kube-api-access-fhz6c\") pod \"dns-default-579gb\" (UID: \"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86\") " pod="openshift-dns/dns-default-579gb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.298210 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.300334 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.300504 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.800481561 +0000 UTC m=+142.949522517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.300667 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.301021 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.801005156 +0000 UTC m=+142.950046102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.302666 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.310309 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.314404 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6hbr\" (UniqueName: \"kubernetes.io/projected/2030e722-71ba-45d8-8d57-07a317e966e2-kube-api-access-s6hbr\") pod \"csi-hostpathplugin-sn98k\" (UID: \"2030e722-71ba-45d8-8d57-07a317e966e2\") " pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.330837 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-sn98k" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.332203 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8kzz\" (UniqueName: \"kubernetes.io/projected/17e43205-39e4-4b98-8d08-4ecd5b1c6d81-kube-api-access-k8kzz\") pod \"kube-storage-version-migrator-operator-b67b599dd-9jgzb\" (UID: \"17e43205-39e4-4b98-8d08-4ecd5b1c6d81\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.337566 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.342267 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-579gb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.348115 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hjt8z" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.353837 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sbbhd" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.361390 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-846vj\" (UniqueName: \"kubernetes.io/projected/a7d578bd-abe7-4012-8f91-04a16dafeb80-kube-api-access-846vj\") pod \"packageserver-d55dfcdfc-8k58l\" (UID: \"a7d578bd-abe7-4012-8f91-04a16dafeb80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.367598 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hs22\" (UniqueName: \"kubernetes.io/projected/561d47d5-3264-46bb-9796-356ab9cfaaf8-kube-api-access-7hs22\") pod \"multus-admission-controller-857f4d67dd-27hxl\" (UID: \"561d47d5-3264-46bb-9796-356ab9cfaaf8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.391086 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwqv2\" (UniqueName: \"kubernetes.io/projected/f8a59726-54f0-4469-b7fc-242607f33e83-kube-api-access-cwqv2\") pod \"marketplace-operator-79b997595-mm77f\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.401503 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.401769 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:34.901755517 +0000 UTC m=+143.050796463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.414576 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg258\" (UniqueName: \"kubernetes.io/projected/0439b8a1-3e02-42a7-92ba-9a7fca3b3636-kube-api-access-pg258\") pod \"package-server-manager-789f6589d5-hqxtv\" (UID: \"0439b8a1-3e02-42a7-92ba-9a7fca3b3636\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.487942 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z8zkl"] Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.502613 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.502957 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.00294613 +0000 UTC m=+143.151987086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.507814 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.536104 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.570942 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.577517 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.586233 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.590905 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.603127 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.603559 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.103544446 +0000 UTC m=+143.252585402 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.707660 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.708163 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.20815204 +0000 UTC m=+143.357192996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.810937 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.811530 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.311514617 +0000 UTC m=+143.460555573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.826086 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" event={"ID":"c50d7dd0-c9f5-4dc6-87df-27c70cf80ed8","Type":"ContainerStarted","Data":"81d0fc95828098ed8ae040f190c48b38e57973416188dc1da30143a0a45aced1"} Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.842518 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk"] Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.849586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" event={"ID":"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4","Type":"ContainerStarted","Data":"70ecefae6506356fdd3625ba8d7e7769ae35c098035f86ff05d260c7140a7c53"} Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.849620 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" event={"ID":"83b28544-0d5c-4c19-87cf-ee6fc37c4ed4","Type":"ContainerStarted","Data":"617bb5da1a5082785412bb61bea5e6c4ffd4bcac77434b0eadc80704699da027"} Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.865677 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" event={"ID":"59a6080f-c880-4307-ae04-090a526e528d","Type":"ContainerStarted","Data":"fa2acd29b6d2c41d02bfe3e440533d87890ef88d3cca5260e886a53e9bc7cd01"} Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.868415 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" event={"ID":"0a8210d9-a926-43ad-8830-6b41e594a60e","Type":"ContainerStarted","Data":"e69f1874d197157493261e73e1797fbd3608cc0dd230412d9d83beb3d78071a7"} Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.889043 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj"] Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.919024 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:34 crc kubenswrapper[4726]: E1123 20:10:34.929412 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.429392691 +0000 UTC m=+143.578433647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.946302 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" event={"ID":"187fba44-db16-4325-9df5-75ab2f9da2d3","Type":"ContainerStarted","Data":"c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660"} Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.947386 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.978796 4726 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-vmvvq container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" start-of-body= Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.978859 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" podUID="187fba44-db16-4325-9df5-75ab2f9da2d3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" Nov 23 20:10:34 crc kubenswrapper[4726]: I1123 20:10:34.993891 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:34.999960 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.019502 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.022137 4726 generic.go:334] "Generic (PLEG): container finished" podID="dd8c701f-32dc-479c-a956-b8cbd83def11" containerID="fac00c2193e5dce8af3db1ee0d0a44dcd8d5eab0f3e83da9ddf50221c272eeb2" exitCode=0 Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.022381 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" event={"ID":"dd8c701f-32dc-479c-a956-b8cbd83def11","Type":"ContainerDied","Data":"fac00c2193e5dce8af3db1ee0d0a44dcd8d5eab0f3e83da9ddf50221c272eeb2"} Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.035838 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.038780 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.538756385 +0000 UTC m=+143.687797341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.047788 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" event={"ID":"002cee03-8433-4e0d-b648-45300fadd459","Type":"ContainerStarted","Data":"c2504d2450558e544794bf13cb4977bf175b1c16362e41eb4ce4b737779c6562"} Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.051655 4726 generic.go:334] "Generic (PLEG): container finished" podID="8dc404f4-e217-4063-bb48-4d06d0802f97" containerID="b5b3468be68ba899a797074b55a1379732bab40ad5d9883c695ebded00cfe2c2" exitCode=0 Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.051703 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" event={"ID":"8dc404f4-e217-4063-bb48-4d06d0802f97","Type":"ContainerDied","Data":"b5b3468be68ba899a797074b55a1379732bab40ad5d9883c695ebded00cfe2c2"} Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.061335 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-4wfvf" event={"ID":"a9efda09-ccd1-47d3-bdc1-b397b2e751f1","Type":"ContainerStarted","Data":"8c614d92167c0cf69d00cfda4a2f7e066a9de4666d638eb2372f90c94287c49c"} Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.063720 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-llmxt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.063759 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-llmxt" podUID="7a3756f7-3ccd-4b4b-a888-eb64139a89a3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.064066 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" event={"ID":"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb","Type":"ContainerStarted","Data":"ac2ee98e78ababf3502ef4e63380463e33f650a40537f7647b9428df401d8b46"} Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.064103 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" event={"ID":"67c2c0c5-6b9c-4f92-a65c-5ce5a41ae6eb","Type":"ContainerStarted","Data":"383a614aee89f111463fe52b3aa756425a7ec2ea16d966c4c789774711757301"} Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.064665 4726 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dnsz8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.064691 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" podUID="81862660-59a0-461e-b7b1-bf53c6e7a5e3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.065045 4726 patch_prober.go:28] interesting pod/console-operator-58897d9998-dvnr5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.065073 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" podUID="e01392f4-eeca-4d54-89eb-2b45c1a807d4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.065602 4726 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7ngz8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.065625 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" podUID="b896681a-7b17-4f20-95a2-f0232ef91f3d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.140860 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.141969 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.641955177 +0000 UTC m=+143.790996133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.166228 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" podStartSLOduration=123.16621209 podStartE2EDuration="2m3.16621209s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:35.12332022 +0000 UTC m=+143.272361176" watchObservedRunningTime="2025-11-23 20:10:35.16621209 +0000 UTC m=+143.315253046" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.242001 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.242425 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.742412299 +0000 UTC m=+143.891453255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.253644 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jsqsq"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.339853 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vfcn6" podStartSLOduration=123.339834452 podStartE2EDuration="2m3.339834452s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:35.298219129 +0000 UTC m=+143.447260085" watchObservedRunningTime="2025-11-23 20:10:35.339834452 +0000 UTC m=+143.488875398" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.340750 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" podStartSLOduration=123.340745119 podStartE2EDuration="2m3.340745119s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:35.340193442 +0000 UTC m=+143.489234408" watchObservedRunningTime="2025-11-23 20:10:35.340745119 +0000 UTC m=+143.489786075" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.344806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.345072 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.845061366 +0000 UTC m=+143.994102322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.451093 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.451449 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:35.951436191 +0000 UTC m=+144.100477147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.495173 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.549851 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.564968 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.565541 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.065526584 +0000 UTC m=+144.214567530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.598920 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.640992 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8phvk" podStartSLOduration=123.640976161 podStartE2EDuration="2m3.640976161s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:35.640557058 +0000 UTC m=+143.789598014" watchObservedRunningTime="2025-11-23 20:10:35.640976161 +0000 UTC m=+143.790017117" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.641403 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9ch2n" podStartSLOduration=124.641397963 podStartE2EDuration="2m4.641397963s" podCreationTimestamp="2025-11-23 20:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:35.616020837 +0000 UTC m=+143.765061793" watchObservedRunningTime="2025-11-23 20:10:35.641397963 +0000 UTC m=+143.790438919" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.669274 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.669647 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.169633443 +0000 UTC m=+144.318674399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.690305 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-579gb"] Nov 23 20:10:35 crc kubenswrapper[4726]: W1123 20:10:35.706874 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c969e63_8584_4613_85be_1d11f3014d13.slice/crio-085437d0a385d3056b90092abe36ddc8eff80b4275626387cf9afa022451f8b0 WatchSource:0}: Error finding container 085437d0a385d3056b90092abe36ddc8eff80b4275626387cf9afa022451f8b0: Status 404 returned error can't find the container with id 085437d0a385d3056b90092abe36ddc8eff80b4275626387cf9afa022451f8b0 Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.708483 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t26mb" podStartSLOduration=123.708469204 podStartE2EDuration="2m3.708469204s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:35.708419252 +0000 UTC m=+143.857460218" watchObservedRunningTime="2025-11-23 20:10:35.708469204 +0000 UTC m=+143.857510160" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.772668 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.773298 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.273286929 +0000 UTC m=+144.422327875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.827860 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sn98k"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.873833 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.874197 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.374151142 +0000 UTC m=+144.523192088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.928195 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" podStartSLOduration=123.9281807 podStartE2EDuration="2m3.9281807s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:35.897656893 +0000 UTC m=+144.046697849" watchObservedRunningTime="2025-11-23 20:10:35.9281807 +0000 UTC m=+144.077221656" Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.962294 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9q98t"] Nov 23 20:10:35 crc kubenswrapper[4726]: I1123 20:10:35.975631 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:35 crc kubenswrapper[4726]: E1123 20:10:35.976026 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.476012736 +0000 UTC m=+144.625053692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.078847 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.079136 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.579119975 +0000 UTC m=+144.728160931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.103017 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.129102 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs"] Nov 23 20:10:36 crc kubenswrapper[4726]: W1123 20:10:36.153543 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea43f27a_45f0_42e3_9c85_360c4e4578cf.slice/crio-e3aaf77da978c0eb332eed555bd23c68987be03e5bd0ac045204eb74dff38a3f WatchSource:0}: Error finding container e3aaf77da978c0eb332eed555bd23c68987be03e5bd0ac045204eb74dff38a3f: Status 404 returned error can't find the container with id e3aaf77da978c0eb332eed555bd23c68987be03e5bd0ac045204eb74dff38a3f Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.155501 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" event={"ID":"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa","Type":"ContainerStarted","Data":"4f46b888ef02ecd54cb8a3d467ee73754c3db5eef8d6b2cb1d06575e603be976"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.170606 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hjt8z"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.184400 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" event={"ID":"8c969e63-8584-4613-85be-1d11f3014d13","Type":"ContainerStarted","Data":"085437d0a385d3056b90092abe36ddc8eff80b4275626387cf9afa022451f8b0"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.185026 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.185365 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.685353767 +0000 UTC m=+144.834394713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.194600 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-q4zwl" podStartSLOduration=124.194584408 podStartE2EDuration="2m4.194584408s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:36.183499193 +0000 UTC m=+144.332540149" watchObservedRunningTime="2025-11-23 20:10:36.194584408 +0000 UTC m=+144.343625364" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.195371 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.236537 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.286998 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g5k42" podStartSLOduration=124.286982063 podStartE2EDuration="2m4.286982063s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:36.28107202 +0000 UTC m=+144.430112976" watchObservedRunningTime="2025-11-23 20:10:36.286982063 +0000 UTC m=+144.436023019" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.287679 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.288023 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.788012574 +0000 UTC m=+144.937053530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.292744 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" event={"ID":"3384ef94-8855-487c-9457-1ff0d7a393c9","Type":"ContainerStarted","Data":"8e020421fce20b2841650c3799e81b362940a4370566982bdf11b65ac54eadea"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.342628 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" event={"ID":"59a6080f-c880-4307-ae04-090a526e528d","Type":"ContainerStarted","Data":"adfd49f24438dc6aeff6e4da2a9c5aee0f5d532a133df323ee26788841f0f701"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.361021 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27hxl"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.364021 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" event={"ID":"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006","Type":"ContainerStarted","Data":"26ff94afb688d72d1b6fb00ac50925a681a1cefec17beea737f1e37d46799801"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.390264 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.392175 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.892162044 +0000 UTC m=+145.041203000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.397115 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-4wfvf" event={"ID":"a9efda09-ccd1-47d3-bdc1-b397b2e751f1","Type":"ContainerStarted","Data":"be5cb1275f2c8820d36e31b3160b81e9ff3078a3eab659e3727408672ffdabf7"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.406604 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" podStartSLOduration=123.406588408 podStartE2EDuration="2m3.406588408s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:36.405324671 +0000 UTC m=+144.554365627" watchObservedRunningTime="2025-11-23 20:10:36.406588408 +0000 UTC m=+144.555629364" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.414312 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.428789 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" event={"ID":"4ff399e8-7c2b-497c-aec8-96383b2f92fb","Type":"ContainerStarted","Data":"272cfb7bee617d83153ad41ee0ebd7efcab446175bed1c7f387f246f272e0a92"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.428979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" event={"ID":"4ff399e8-7c2b-497c-aec8-96383b2f92fb","Type":"ContainerStarted","Data":"2d16c8e3191844875adf56bffeed4c26e2370c55dff9bff0c8f8f0605800d17d"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.463294 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sbbhd" event={"ID":"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a","Type":"ContainerStarted","Data":"eeb6caafeaa40f11a20602ac861a1015dbe6864fb6ba694f45bbbffab5199927"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.463357 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sbbhd" event={"ID":"0f95577a-bb59-4dbb-9bb3-8d4f18e5c89a","Type":"ContainerStarted","Data":"9cce6426e01e76d3c3f741645f6cdd6cec38e8055e64c6017f9fa3f128035384"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.491923 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.493611 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:36.993593894 +0000 UTC m=+145.142634850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: W1123 20:10:36.510218 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf62d5e1f_4d77_4474_aa97_6966ad32e2c2.slice/crio-d23b147f458e64a32ccb1f611fa9b0557d7b39a2c760dae0e3ea5edaa6165760 WatchSource:0}: Error finding container d23b147f458e64a32ccb1f611fa9b0557d7b39a2c760dae0e3ea5edaa6165760: Status 404 returned error can't find the container with id d23b147f458e64a32ccb1f611fa9b0557d7b39a2c760dae0e3ea5edaa6165760 Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.539691 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.566583 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.573628 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sn98k" event={"ID":"2030e722-71ba-45d8-8d57-07a317e966e2","Type":"ContainerStarted","Data":"c374faebd099b2c8b1c467d1fe89f8b1f2c760b0e05269c66ab5fe2dba1627b7"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.582330 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mm77f"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.584355 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" event={"ID":"581ff47c-292f-4901-8ee4-124258f07696","Type":"ContainerStarted","Data":"66339b0b99785b5f6dc8c9bb32340d6f3592ad5a4d970834914c98b41b573eb9"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.585424 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-579gb" event={"ID":"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86","Type":"ContainerStarted","Data":"5894341685fa82c517d8d342cd3d063d54ef533dcd9b189796fb18efead4ee7b"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.586890 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w77x8" podStartSLOduration=124.586857805 podStartE2EDuration="2m4.586857805s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:36.58464984 +0000 UTC m=+144.733690806" watchObservedRunningTime="2025-11-23 20:10:36.586857805 +0000 UTC m=+144.735898761" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.587590 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" event={"ID":"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1","Type":"ContainerStarted","Data":"c543a77948ee2d938d55c6d1d9b200239dc06a54c0672bb8c2de13dd3c88042f"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.587621 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" event={"ID":"8e41e7f4-8c3f-431c-aa5a-fa79132ceef1","Type":"ContainerStarted","Data":"4da76c9e48b149ca9ac1984926167b8c0a9e7817dfc19ba8608a60a2f12c6078"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.587932 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.594067 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.595132 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.095118488 +0000 UTC m=+145.244159444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.613117 4726 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gwlrk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.613165 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" podUID="8e41e7f4-8c3f-431c-aa5a-fa79132ceef1" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.654606 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l"] Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.654644 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" event={"ID":"b4e6d66f-e455-4f27-9f46-a03b2f20cb51","Type":"ContainerStarted","Data":"05d71987bbb755fe85653f227d958e1b5164f47a35d182b8962d1630dfa6f47f"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.654660 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" event={"ID":"b4e6d66f-e455-4f27-9f46-a03b2f20cb51","Type":"ContainerStarted","Data":"d820bb4190c00e060831dae956e55f65d4368f57d45b76f55f8c198cb9a8b6bd"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.695403 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.695681 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.195664882 +0000 UTC m=+145.344705838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.695772 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.696894 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.196885168 +0000 UTC m=+145.345926124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.721688 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" event={"ID":"36126794-ca3b-42fa-b977-adcc62a6902e","Type":"ContainerStarted","Data":"ba9c5159625b3c52a790469f7fbd5122e8ffa282207b59277ca95134d0a91750"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.723934 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" podStartSLOduration=123.723917082 podStartE2EDuration="2m3.723917082s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:36.72350875 +0000 UTC m=+144.872549706" watchObservedRunningTime="2025-11-23 20:10:36.723917082 +0000 UTC m=+144.872958038" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.780570 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vfgwr" event={"ID":"002cee03-8433-4e0d-b648-45300fadd459","Type":"ContainerStarted","Data":"398fc9edcd0ddcfce88aca3df7ca6c14509ccf3f6b4511e06d3900135557ec34"} Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.782307 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-llmxt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.782659 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-llmxt" podUID="7a3756f7-3ccd-4b4b-a888-eb64139a89a3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.797242 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.797590 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.297570357 +0000 UTC m=+145.446611313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.831100 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.833449 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-llmxt" podStartSLOduration=124.83342684 podStartE2EDuration="2m4.83342684s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:36.780473894 +0000 UTC m=+144.929514850" watchObservedRunningTime="2025-11-23 20:10:36.83342684 +0000 UTC m=+144.982467796" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.866539 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:10:36 crc kubenswrapper[4726]: I1123 20:10:36.916850 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:36 crc kubenswrapper[4726]: E1123 20:10:36.924987 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.417571983 +0000 UTC m=+145.566612939 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.020459 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.020724 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.520705213 +0000 UTC m=+145.669746169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.020775 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.021117 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.521111305 +0000 UTC m=+145.670152261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.040375 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" podStartSLOduration=125.040357051 podStartE2EDuration="2m5.040357051s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:37.010422241 +0000 UTC m=+145.159463207" watchObservedRunningTime="2025-11-23 20:10:37.040357051 +0000 UTC m=+145.189398007" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.082016 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" podStartSLOduration=125.082003115 podStartE2EDuration="2m5.082003115s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:37.078915384 +0000 UTC m=+145.227956360" watchObservedRunningTime="2025-11-23 20:10:37.082003115 +0000 UTC m=+145.231044071" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.108610 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hlvv6" podStartSLOduration=124.108594516 podStartE2EDuration="2m4.108594516s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:37.106206926 +0000 UTC m=+145.255247882" watchObservedRunningTime="2025-11-23 20:10:37.108594516 +0000 UTC m=+145.257635472" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.121305 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.121581 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.621570067 +0000 UTC m=+145.770611023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.124809 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-dvnr5" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.222556 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.222884 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.722860743 +0000 UTC m=+145.871901699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.264046 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.267679 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:37 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:37 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:37 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.267712 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.323343 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.323626 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.823612284 +0000 UTC m=+145.972653230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.392509 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" podStartSLOduration=124.392496968 podStartE2EDuration="2m4.392496968s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:37.390214541 +0000 UTC m=+145.539255487" watchObservedRunningTime="2025-11-23 20:10:37.392496968 +0000 UTC m=+145.541537924" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.427383 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.427763 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:37.927746084 +0000 UTC m=+146.076787040 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.506526 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-sbbhd" podStartSLOduration=6.506510598 podStartE2EDuration="6.506510598s" podCreationTimestamp="2025-11-23 20:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:37.505839369 +0000 UTC m=+145.654880325" watchObservedRunningTime="2025-11-23 20:10:37.506510598 +0000 UTC m=+145.655551554" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.507189 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-4wfvf" podStartSLOduration=125.507183318 podStartE2EDuration="2m5.507183318s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:37.435188883 +0000 UTC m=+145.584229829" watchObservedRunningTime="2025-11-23 20:10:37.507183318 +0000 UTC m=+145.656224274" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.528619 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.528985 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.028971838 +0000 UTC m=+146.178012794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.634275 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.634605 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.134594342 +0000 UTC m=+146.283635298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.641480 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.641796 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.735600 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.736282 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.236247609 +0000 UTC m=+146.385288565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.780182 4726 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-vmvvq container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.17:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.780244 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" podUID="187fba44-db16-4325-9df5-75ab2f9da2d3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.17:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.807163 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" event={"ID":"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa","Type":"ContainerStarted","Data":"2b1fea63a20f4253882bf1e7a1a4022c533207dd3b202c943c0ffc24ed3b0659"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.807214 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" event={"ID":"8cd65cd8-b98b-41df-8ac3-ec8e6a83b0fa","Type":"ContainerStarted","Data":"f24593e4fd6a4372b39afd44d2428040797165d633231bbb2e6ca76c5ef43f1f"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.828456 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" event={"ID":"b4e6d66f-e455-4f27-9f46-a03b2f20cb51","Type":"ContainerStarted","Data":"6f0d8dee9569ae0d19fa60cad5996afea7b6d9b5bb3e391340bef7764bf90ceb"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.840592 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.840910 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.340897603 +0000 UTC m=+146.489938569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.845916 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" event={"ID":"a7d578bd-abe7-4012-8f91-04a16dafeb80","Type":"ContainerStarted","Data":"7a1155273c4c2f77d4b81dbae797fbe8eecea3f956a2dfaabc3b68135824d1de"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.857922 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" event={"ID":"581ff47c-292f-4901-8ee4-124258f07696","Type":"ContainerStarted","Data":"8b3a3784b81a2dd97ce7012a03d7326e2616c9fac8a43b67a277f79714046ff6"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.866230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hjt8z" event={"ID":"f62d5e1f-4d77-4474-aa97-6966ad32e2c2","Type":"ContainerStarted","Data":"d23b147f458e64a32ccb1f611fa9b0557d7b39a2c760dae0e3ea5edaa6165760"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.872374 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" event={"ID":"36126794-ca3b-42fa-b977-adcc62a6902e","Type":"ContainerStarted","Data":"0adae51ad351b692d825cc3d3f8e552d1c45b481440d2cfe51255637fe2a3ba6"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.875766 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" event={"ID":"2077c584-5d3d-48bc-9c39-2e5fc941cc38","Type":"ContainerStarted","Data":"04b0f779695e4212a3b6f4786ecac438c9fe31fcb2636d12dc9eeb646cba2744"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.881633 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" event={"ID":"dd8c701f-32dc-479c-a956-b8cbd83def11","Type":"ContainerStarted","Data":"80128ea7983c776c6ff3d87151b94abed18b8a480d21f9ab53ccfecd00a5c49e"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.889010 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" event={"ID":"8dc404f4-e217-4063-bb48-4d06d0802f97","Type":"ContainerStarted","Data":"4807f5a0eac5c5f98e05b727e116282f89df2278d236d323aaa54010f887b7df"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.893416 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" event={"ID":"561d47d5-3264-46bb-9796-356ab9cfaaf8","Type":"ContainerStarted","Data":"135e18925746d5a495cd31eca48d72e16b2b9271b8af274b1bdb2190edda2bbc"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.898262 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m2fhd" podStartSLOduration=124.898247658 podStartE2EDuration="2m4.898247658s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:37.873604424 +0000 UTC m=+146.022645390" watchObservedRunningTime="2025-11-23 20:10:37.898247658 +0000 UTC m=+146.047288614" Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.932126 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" event={"ID":"ea43f27a-45f0-42e3-9c85-360c4e4578cf","Type":"ContainerStarted","Data":"6cbeba3da27095ddcef6e3982482c343754466b16341fc4bb6c0042853c05227"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.932167 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" event={"ID":"ea43f27a-45f0-42e3-9c85-360c4e4578cf","Type":"ContainerStarted","Data":"e3aaf77da978c0eb332eed555bd23c68987be03e5bd0ac045204eb74dff38a3f"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.942799 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.949498 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" event={"ID":"8ae6fd55-a974-4701-a37f-b263979a310c","Type":"ContainerStarted","Data":"1db4228f2f9b702e8e146ea516a92ee16bc60164a5f4bcd75b0c380061670c82"} Nov 23 20:10:37 crc kubenswrapper[4726]: E1123 20:10:37.950328 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.450305948 +0000 UTC m=+146.599346904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.986737 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" event={"ID":"f8a59726-54f0-4469-b7fc-242607f33e83","Type":"ContainerStarted","Data":"49d0b24032c0976f1a239fd9d4023798ecf0dd4cd428cd11e8b5562e1cb7aaf8"} Nov 23 20:10:37 crc kubenswrapper[4726]: I1123 20:10:37.987797 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.001298 4726 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mm77f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.001349 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.035654 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.035966 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.038624 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" event={"ID":"8c969e63-8584-4613-85be-1d11f3014d13","Type":"ContainerStarted","Data":"a1936471dbabc26a0a6359a5ef2e5d10d2bb1338d573d418a8e49cb44f3567e5"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.044401 4726 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-z4rlj container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.044443 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" podUID="dd8c701f-32dc-479c-a956-b8cbd83def11" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.044653 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.046461 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.546449313 +0000 UTC m=+146.695490269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.046590 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" event={"ID":"0439b8a1-3e02-42a7-92ba-9a7fca3b3636","Type":"ContainerStarted","Data":"efed67c3dc074d2fb12586fe7e936277a413a0d33fcd830d7a926585515e33bf"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.081347 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" event={"ID":"17e43205-39e4-4b98-8d08-4ecd5b1c6d81","Type":"ContainerStarted","Data":"41545eea0d9492314820ed92f05d610731542d2b949ea0be0ba537b8efac1a78"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.081390 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" event={"ID":"17e43205-39e4-4b98-8d08-4ecd5b1c6d81","Type":"ContainerStarted","Data":"9a351abf18c4503ba2bee37566a5043c1c60b0bd4b025c93f03af3c83d5e52d4"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.088332 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" event={"ID":"e7ac13f8-5f8d-4eaa-acd1-93f5e15b4006","Type":"ContainerStarted","Data":"9a012d41f5170ff4ba17babf9ec0e742eadc9376cb22690011973cda3eb7e0b6"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.101959 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-579gb" event={"ID":"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86","Type":"ContainerStarted","Data":"2ddd4e6be6bb9c011b14179e7da5a68f55c7b5f4d7bd32c3dd8666377ed468bf"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.105464 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" event={"ID":"0a8210d9-a926-43ad-8830-6b41e594a60e","Type":"ContainerStarted","Data":"40dc4e6f40f849cf49b04bfe33e08963485620e78da5ec2ce9f6c94420a40445"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.117612 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cv7p8" podStartSLOduration=125.117593684 podStartE2EDuration="2m5.117593684s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.078703181 +0000 UTC m=+146.227744137" watchObservedRunningTime="2025-11-23 20:10:38.117593684 +0000 UTC m=+146.266634640" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.119003 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-jsqsq" podStartSLOduration=126.118996505 podStartE2EDuration="2m6.118996505s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:37.967992198 +0000 UTC m=+146.117033174" watchObservedRunningTime="2025-11-23 20:10:38.118996505 +0000 UTC m=+146.268037461" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.153708 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hxw6z" podStartSLOduration=126.153688294 podStartE2EDuration="2m6.153688294s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.134214592 +0000 UTC m=+146.283255558" watchObservedRunningTime="2025-11-23 20:10:38.153688294 +0000 UTC m=+146.302729250" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.155526 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.156502 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.656491077 +0000 UTC m=+146.805532033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.173004 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" event={"ID":"178c9142-3c12-40db-a898-81f61fd478f8","Type":"ContainerStarted","Data":"96a58cffd64a535c29807fd91a939cd1d1b632cbacad11de84f9ae577d502209"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.173829 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.187519 4726 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6jlhs container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.187577 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" podUID="178c9142-3c12-40db-a898-81f61fd478f8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.211906 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" event={"ID":"10063588-a897-402a-b308-7e892b096b33","Type":"ContainerStarted","Data":"6af5d5ab0c2c02ca1c23fcad69631cbe19a6d1fdde06d017538c9b99792b94c0"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.213397 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" podStartSLOduration=126.213380378 podStartE2EDuration="2m6.213380378s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.211699069 +0000 UTC m=+146.360740025" watchObservedRunningTime="2025-11-23 20:10:38.213380378 +0000 UTC m=+146.362421334" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.259425 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" event={"ID":"3384ef94-8855-487c-9457-1ff0d7a393c9","Type":"ContainerStarted","Data":"85f68d1e1733b625d4725870b0c281b597a013a035aa5cde79063c6536f48461"} Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.261666 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.261923 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.761912664 +0000 UTC m=+146.910953620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.273585 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:38 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:38 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:38 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.273664 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.285102 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.301129 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gwlrk" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.314084 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9q98t" podStartSLOduration=125.314073417 podStartE2EDuration="2m5.314073417s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.260922405 +0000 UTC m=+146.409963371" watchObservedRunningTime="2025-11-23 20:10:38.314073417 +0000 UTC m=+146.463114373" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.370737 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.372726 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.87270446 +0000 UTC m=+147.021745416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.471845 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mcmsj" podStartSLOduration=126.471828233 podStartE2EDuration="2m6.471828233s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.320008441 +0000 UTC m=+146.469049417" watchObservedRunningTime="2025-11-23 20:10:38.471828233 +0000 UTC m=+146.620869179" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.475556 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.475857 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:38.975845831 +0000 UTC m=+147.124886787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.576887 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.577307 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.077270821 +0000 UTC m=+147.226311777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.596267 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" podStartSLOduration=125.596248459 podStartE2EDuration="2m5.596248459s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.47616101 +0000 UTC m=+146.625201976" watchObservedRunningTime="2025-11-23 20:10:38.596248459 +0000 UTC m=+146.745289415" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.643433 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" podStartSLOduration=125.643416605 podStartE2EDuration="2m5.643416605s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.598042011 +0000 UTC m=+146.747082967" watchObservedRunningTime="2025-11-23 20:10:38.643416605 +0000 UTC m=+146.792457551" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.644805 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9jgzb" podStartSLOduration=125.644800295 podStartE2EDuration="2m5.644800295s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.643276871 +0000 UTC m=+146.792317827" watchObservedRunningTime="2025-11-23 20:10:38.644800295 +0000 UTC m=+146.793841241" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.679360 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.679590 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.179576777 +0000 UTC m=+147.328617733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.703913 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dvndr" podStartSLOduration=126.703895192 podStartE2EDuration="2m6.703895192s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.703085638 +0000 UTC m=+146.852126604" watchObservedRunningTime="2025-11-23 20:10:38.703895192 +0000 UTC m=+146.852936148" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.782369 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.782692 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.282667266 +0000 UTC m=+147.431708212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.884102 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.884409 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.384382065 +0000 UTC m=+147.533423021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.887418 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" podStartSLOduration=125.887408124 podStartE2EDuration="2m5.887408124s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:38.742363292 +0000 UTC m=+146.891404248" watchObservedRunningTime="2025-11-23 20:10:38.887408124 +0000 UTC m=+147.036449080" Nov 23 20:10:38 crc kubenswrapper[4726]: I1123 20:10:38.985999 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:38 crc kubenswrapper[4726]: E1123 20:10:38.986331 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.486316091 +0000 UTC m=+147.635357047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.043723 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.043785 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.087696 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.088112 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.588093881 +0000 UTC m=+147.737134837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.188709 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.188923 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.688897433 +0000 UTC m=+147.837938389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.265708 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-579gb" event={"ID":"b66e5a55-9c4e-4383-9d4f-6961fbd5ab86","Type":"ContainerStarted","Data":"20169752970cf0da76fe038a3631d500982007c72436c981790f4f00ba3246b9"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.266152 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-579gb" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.267121 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" event={"ID":"0439b8a1-3e02-42a7-92ba-9a7fca3b3636","Type":"ContainerStarted","Data":"667ffc3cc6144d091a0d3dcfcc0190d2233e33a3c80e8ad2f6e4573364957050"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.267161 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" event={"ID":"0439b8a1-3e02-42a7-92ba-9a7fca3b3636","Type":"ContainerStarted","Data":"7eb97a5536f74dbf6ee1ebde235a24d87593924b92f474594ff58aebe9a88549"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.267315 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.268537 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" event={"ID":"a7d578bd-abe7-4012-8f91-04a16dafeb80","Type":"ContainerStarted","Data":"794d499b8ec226597fa2fd8eb42725c3f3de24422117095f3576e18969a6e4a2"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.269158 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.270210 4726 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8k58l container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.270244 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" podUID="a7d578bd-abe7-4012-8f91-04a16dafeb80" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.270638 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" event={"ID":"0a8210d9-a926-43ad-8830-6b41e594a60e","Type":"ContainerStarted","Data":"4ff55e5e5a70c7ae1ec0b54c6999e7ebee29c244eee65e767185e165a6c9b9bd"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.271770 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" event={"ID":"178c9142-3c12-40db-a898-81f61fd478f8","Type":"ContainerStarted","Data":"ac2fad2ea041d609dd01f2c1e5fcdbc742fa0134c56b75d440175ee9457eb250"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.272509 4726 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6jlhs container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.272533 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" podUID="178c9142-3c12-40db-a898-81f61fd478f8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.274376 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hjt8z" event={"ID":"f62d5e1f-4d77-4474-aa97-6966ad32e2c2","Type":"ContainerStarted","Data":"5e18f9535134caae072c79c4d8e384031aad7165c957f07d946063935f84f0d1"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.275620 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" event={"ID":"2077c584-5d3d-48bc-9c39-2e5fc941cc38","Type":"ContainerStarted","Data":"2b3ff5233d2350970e2012e257eb3c7769e140c9adcc68e990d40aca38a75e3c"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.276119 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:39 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:39 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:39 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.276167 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.278404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" event={"ID":"561d47d5-3264-46bb-9796-356ab9cfaaf8","Type":"ContainerStarted","Data":"016347f12fc3be7d028ef20d931299a15ef8c43ed0b252934ef4f1b7dc70aebb"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.278430 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" event={"ID":"561d47d5-3264-46bb-9796-356ab9cfaaf8","Type":"ContainerStarted","Data":"f3d595a21df01c311b4af4c16734e9e8d04d43e87217c64b16ecca71c4b0224f"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.280222 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" event={"ID":"8ae6fd55-a974-4701-a37f-b263979a310c","Type":"ContainerStarted","Data":"1a1f4a8baa59589d6be56a612878b6a3bfb780ed98ec7b6f1cafd1baa9c16fdd"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.280265 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" event={"ID":"8ae6fd55-a974-4701-a37f-b263979a310c","Type":"ContainerStarted","Data":"5a77afd08c7cd492c15d37f0c09d7d041891685a4b90d6fc5a04ef9ee4ddd9d0"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.281459 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sn98k" event={"ID":"2030e722-71ba-45d8-8d57-07a317e966e2","Type":"ContainerStarted","Data":"f69b601a9d9dd666bbf80ece7447be3bcb24a7f46316a1067d36c4dd7db6625e"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.282413 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" event={"ID":"f8a59726-54f0-4469-b7fc-242607f33e83","Type":"ContainerStarted","Data":"b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.283143 4726 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mm77f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.283179 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.288714 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" event={"ID":"10063588-a897-402a-b308-7e892b096b33","Type":"ContainerStarted","Data":"e5738c9270a51d849dda4abb9749c4852d86917ce70e13e09c261ff81c26166b"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.288751 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" event={"ID":"10063588-a897-402a-b308-7e892b096b33","Type":"ContainerStarted","Data":"1128d606e2c4add3a826e653b9c708ed7f6fc8202a932cde14e4f23c022cac93"} Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.290650 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.291063 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.791041185 +0000 UTC m=+147.940082141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.292153 4726 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-f8fmr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.292213 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" podUID="8dc404f4-e217-4063-bb48-4d06d0802f97" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.363580 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-579gb" podStartSLOduration=8.363563936 podStartE2EDuration="8.363563936s" podCreationTimestamp="2025-11-23 20:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.30823228 +0000 UTC m=+147.457273246" watchObservedRunningTime="2025-11-23 20:10:39.363563936 +0000 UTC m=+147.512604892" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.365453 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-27hxl" podStartSLOduration=126.365448571 podStartE2EDuration="2m6.365448571s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.360036692 +0000 UTC m=+147.509077648" watchObservedRunningTime="2025-11-23 20:10:39.365448571 +0000 UTC m=+147.514489527" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.394796 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.395005 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.894974969 +0000 UTC m=+148.044015925 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.395108 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.395347 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.395605 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.395666 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.395766 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.403732 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.903713026 +0000 UTC m=+148.052753982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.413536 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.420693 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.420800 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.432053 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.445474 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-z8zkl" podStartSLOduration=127.445460122 podStartE2EDuration="2m7.445460122s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.394213196 +0000 UTC m=+147.543254152" watchObservedRunningTime="2025-11-23 20:10:39.445460122 +0000 UTC m=+147.594501068" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.497256 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.497439 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.997414489 +0000 UTC m=+148.146455445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.497751 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.498068 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:39.998054878 +0000 UTC m=+148.147095834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.504651 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.510172 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hjt8z" podStartSLOduration=8.510156333 podStartE2EDuration="8.510156333s" podCreationTimestamp="2025-11-23 20:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.44641229 +0000 UTC m=+147.595453246" watchObservedRunningTime="2025-11-23 20:10:39.510156333 +0000 UTC m=+147.659197289" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.511662 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ks9vb" podStartSLOduration=126.511657058 podStartE2EDuration="2m6.511657058s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.508880026 +0000 UTC m=+147.657920982" watchObservedRunningTime="2025-11-23 20:10:39.511657058 +0000 UTC m=+147.660698014" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.518983 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.527781 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.602412 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.602774 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.102758944 +0000 UTC m=+148.251799900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.682178 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qs2w2" podStartSLOduration=126.682158068 podStartE2EDuration="2m6.682158068s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.586808706 +0000 UTC m=+147.735849662" watchObservedRunningTime="2025-11-23 20:10:39.682158068 +0000 UTC m=+147.831199024" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.704152 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.704429 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.204416442 +0000 UTC m=+148.353457398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.725625 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" podStartSLOduration=126.725607664 podStartE2EDuration="2m6.725607664s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.683116166 +0000 UTC m=+147.832157122" watchObservedRunningTime="2025-11-23 20:10:39.725607664 +0000 UTC m=+147.874648620" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.795651 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" podStartSLOduration=126.795632842 podStartE2EDuration="2m6.795632842s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.751790594 +0000 UTC m=+147.900831560" watchObservedRunningTime="2025-11-23 20:10:39.795632842 +0000 UTC m=+147.944673798" Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.805308 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.805599 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.305585244 +0000 UTC m=+148.454626200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:39 crc kubenswrapper[4726]: I1123 20:10:39.906682 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:39 crc kubenswrapper[4726]: E1123 20:10:39.907074 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.407057946 +0000 UTC m=+148.556098902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.016355 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.016989 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.516974546 +0000 UTC m=+148.666015502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.117510 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.117882 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.61785174 +0000 UTC m=+148.766892696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.218084 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.218448 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.718433156 +0000 UTC m=+148.867474112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.269315 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:40 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:40 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:40 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.269367 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.295783 4726 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mm77f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.295824 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.324961 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.325386 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.825372968 +0000 UTC m=+148.974413924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.387110 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6jlhs" Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.432303 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.432406 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.932384723 +0000 UTC m=+149.081425679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.432855 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.434544 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:40.934532586 +0000 UTC m=+149.083573542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.472954 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-frgld" podStartSLOduration=127.472936494 podStartE2EDuration="2m7.472936494s" podCreationTimestamp="2025-11-23 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:39.805751649 +0000 UTC m=+147.954792605" watchObservedRunningTime="2025-11-23 20:10:40.472936494 +0000 UTC m=+148.621977450" Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.534586 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.535013 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.034997918 +0000 UTC m=+149.184038874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.636124 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.636404 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.136392877 +0000 UTC m=+149.285433833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.737055 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.737423 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.237408796 +0000 UTC m=+149.386449752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.838466 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.838790 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.338779014 +0000 UTC m=+149.487819970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.874361 4726 patch_prober.go:28] interesting pod/apiserver-76f77b778f-m5z9k container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]log ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]etcd ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/generic-apiserver-start-informers ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/max-in-flight-filter ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 23 20:10:40 crc kubenswrapper[4726]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 23 20:10:40 crc kubenswrapper[4726]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/project.openshift.io-projectcache ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-startinformers ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 23 20:10:40 crc kubenswrapper[4726]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 23 20:10:40 crc kubenswrapper[4726]: livez check failed Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.874419 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" podUID="59a6080f-c880-4307-ae04-090a526e528d" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.940160 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.940402 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.44037535 +0000 UTC m=+149.589416306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: E1123 20:10:40.941386 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.441373859 +0000 UTC m=+149.590414805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:40 crc kubenswrapper[4726]: I1123 20:10:40.941517 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:41 crc kubenswrapper[4726]: W1123 20:10:41.034132 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-df2db5eede7d4f09cceb6a35d49fe3c5c402022a4814d5fc55c99ed9211b04a4 WatchSource:0}: Error finding container df2db5eede7d4f09cceb6a35d49fe3c5c402022a4814d5fc55c99ed9211b04a4: Status 404 returned error can't find the container with id df2db5eede7d4f09cceb6a35d49fe3c5c402022a4814d5fc55c99ed9211b04a4 Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.042141 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.042438 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.542424078 +0000 UTC m=+149.691465034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.143585 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.143989 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.643973012 +0000 UTC m=+149.793013968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.249289 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.249600 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.749583966 +0000 UTC m=+149.898624922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.266490 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:41 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:41 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:41 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.266538 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.296920 4726 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8k58l container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.296964 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" podUID="a7d578bd-abe7-4012-8f91-04a16dafeb80" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.327023 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sn98k" event={"ID":"2030e722-71ba-45d8-8d57-07a317e966e2","Type":"ContainerStarted","Data":"bee31c2d932b05a27de895e6fce818e19da265837efda0a47f28d6f5dd4bbd00"} Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.335637 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"54947d1eac1c3b5b5d3df169a3ec9b68ee7907e420ddf0aef0ee80ea2f1e657b"} Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.335685 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c7061d1cc172638023876b1ead11b32f05597956767e7324c5cc1ba1ca319ac3"} Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.341278 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"df2db5eede7d4f09cceb6a35d49fe3c5c402022a4814d5fc55c99ed9211b04a4"} Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.347926 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5682c980e7f71367a35721dfbbcb9aa88af36e072f973727b4d9d582f4b4dd4a"} Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.350388 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.350644 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.850632835 +0000 UTC m=+149.999673791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.452187 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.453290 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:41.95327143 +0000 UTC m=+150.102312386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.553750 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.554082 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.054068192 +0000 UTC m=+150.203109148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.655169 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.655538 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.155523143 +0000 UTC m=+150.304564099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.757086 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.257076737 +0000 UTC m=+150.406117693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.757353 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.858597 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.858804 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.358774375 +0000 UTC m=+150.507815331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.858856 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.859306 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.359296641 +0000 UTC m=+150.508337597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.959955 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:41 crc kubenswrapper[4726]: E1123 20:10:41.960242 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.460228366 +0000 UTC m=+150.609269322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.994306 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t7csf"] Nov 23 20:10:41 crc kubenswrapper[4726]: I1123 20:10:41.995167 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.000077 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.045273 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7csf"] Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.061076 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-catalog-content\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.061145 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.061167 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlwm6\" (UniqueName: \"kubernetes.io/projected/15f81f72-746c-479f-b5b0-ba9ff588717a-kube-api-access-wlwm6\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.061186 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-utilities\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.061439 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.56142772 +0000 UTC m=+150.710468676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.162586 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.162750 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlwm6\" (UniqueName: \"kubernetes.io/projected/15f81f72-746c-479f-b5b0-ba9ff588717a-kube-api-access-wlwm6\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.162771 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.662746057 +0000 UTC m=+150.811787013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.162815 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-utilities\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.163000 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-catalog-content\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.163680 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-utilities\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.163711 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-catalog-content\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.196164 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h4jxc"] Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.196832 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlwm6\" (UniqueName: \"kubernetes.io/projected/15f81f72-746c-479f-b5b0-ba9ff588717a-kube-api-access-wlwm6\") pod \"community-operators-t7csf\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.196989 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.203173 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.235072 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h4jxc"] Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.235208 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f8fmr" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.264351 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.264412 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stgr5\" (UniqueName: \"kubernetes.io/projected/f1364754-8bfa-45e8-8520-893acdb2837e-kube-api-access-stgr5\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.264441 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-utilities\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.264459 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-catalog-content\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.264726 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.764714554 +0000 UTC m=+150.913755510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.269163 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:42 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:42 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:42 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.269201 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.308130 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.349130 4726 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8k58l container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.349189 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" podUID="a7d578bd-abe7-4012-8f91-04a16dafeb80" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.365448 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.365680 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stgr5\" (UniqueName: \"kubernetes.io/projected/f1364754-8bfa-45e8-8520-893acdb2837e-kube-api-access-stgr5\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.365726 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-utilities\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.365744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-catalog-content\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.366468 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.866454153 +0000 UTC m=+151.015495109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.367227 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-utilities\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.367429 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-catalog-content\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.388347 4726 generic.go:334] "Generic (PLEG): container finished" podID="8c969e63-8584-4613-85be-1d11f3014d13" containerID="a1936471dbabc26a0a6359a5ef2e5d10d2bb1338d573d418a8e49cb44f3567e5" exitCode=0 Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.388428 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" event={"ID":"8c969e63-8584-4613-85be-1d11f3014d13","Type":"ContainerDied","Data":"a1936471dbabc26a0a6359a5ef2e5d10d2bb1338d573d418a8e49cb44f3567e5"} Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.405725 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s5ddf"] Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.406586 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.415680 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stgr5\" (UniqueName: \"kubernetes.io/projected/f1364754-8bfa-45e8-8520-893acdb2837e-kube-api-access-stgr5\") pod \"certified-operators-h4jxc\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.456142 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s5ddf"] Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.456902 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sn98k" event={"ID":"2030e722-71ba-45d8-8d57-07a317e966e2","Type":"ContainerStarted","Data":"40bbd1414096f14bd4aebb5c9c26bff0faca82fd15eaf4ab2a715e24a3cea089"} Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.468168 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs7gr\" (UniqueName: \"kubernetes.io/projected/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-kube-api-access-vs7gr\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.468250 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-utilities\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.468286 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.468333 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-catalog-content\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.468686 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:42.968672567 +0000 UTC m=+151.117713533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.487799 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f40c95d3820852d975696b785859903320c498c699a0992369bb964d156d3663"} Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.507040 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7b35e1313dfbb6a9d11f82cf31a1745261047fd21476250c6c1a5ab27c603941"} Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.507830 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.535179 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.571344 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.571481 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-utilities\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.571546 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-catalog-content\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.571588 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs7gr\" (UniqueName: \"kubernetes.io/projected/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-kube-api-access-vs7gr\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.571730 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.071715935 +0000 UTC m=+151.220756891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.572531 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-utilities\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.573569 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-catalog-content\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.611837 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cn57x"] Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.613325 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.634606 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs7gr\" (UniqueName: \"kubernetes.io/projected/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-kube-api-access-vs7gr\") pod \"community-operators-s5ddf\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.647127 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cn57x"] Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.651305 4726 patch_prober.go:28] interesting pod/apiserver-76f77b778f-m5z9k container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]log ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]etcd ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/generic-apiserver-start-informers ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/max-in-flight-filter ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 23 20:10:42 crc kubenswrapper[4726]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/project.openshift.io-projectcache ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-startinformers ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 23 20:10:42 crc kubenswrapper[4726]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 23 20:10:42 crc kubenswrapper[4726]: livez check failed Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.651363 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" podUID="59a6080f-c880-4307-ae04-090a526e528d" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.683805 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-utilities\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.684100 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.684139 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-catalog-content\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.684209 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m82sj\" (UniqueName: \"kubernetes.io/projected/54adbeda-3036-4850-b945-5416bc4ed456-kube-api-access-m82sj\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.684737 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.184708455 +0000 UTC m=+151.333749411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.718063 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.785347 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.785518 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-utilities\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.785577 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-catalog-content\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.785621 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m82sj\" (UniqueName: \"kubernetes.io/projected/54adbeda-3036-4850-b945-5416bc4ed456-kube-api-access-m82sj\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.785948 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.285934269 +0000 UTC m=+151.434975235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.786120 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-utilities\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.786150 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-catalog-content\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.818696 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-llmxt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.818748 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-llmxt" podUID="7a3756f7-3ccd-4b4b-a888-eb64139a89a3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.818924 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-llmxt container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.818969 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-llmxt" podUID="7a3756f7-3ccd-4b4b-a888-eb64139a89a3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.824620 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m82sj\" (UniqueName: \"kubernetes.io/projected/54adbeda-3036-4850-b945-5416bc4ed456-kube-api-access-m82sj\") pod \"certified-operators-cn57x\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.872264 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.872299 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.873534 4726 patch_prober.go:28] interesting pod/console-f9d7485db-q4zwl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.873576 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-q4zwl" podUID="7445d8ee-ffb1-4610-ae2d-cb511edf1525" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.889472 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.889808 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.389796082 +0000 UTC m=+151.538837028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.942658 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:10:42 crc kubenswrapper[4726]: I1123 20:10:42.991388 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:42 crc kubenswrapper[4726]: E1123 20:10:42.992523 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.49250377 +0000 UTC m=+151.641544726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.026719 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.033459 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-z4rlj" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.094773 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:43 crc kubenswrapper[4726]: E1123 20:10:43.095960 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.595943999 +0000 UTC m=+151.744984955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.164395 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7csf"] Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.198984 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:43 crc kubenswrapper[4726]: E1123 20:10:43.199392 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.699372468 +0000 UTC m=+151.848413434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.234788 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h4jxc"] Nov 23 20:10:43 crc kubenswrapper[4726]: W1123 20:10:43.255077 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1364754_8bfa_45e8_8520_893acdb2837e.slice/crio-dce19ccdb807c181cb4d618ce2ead2bc30290cb5137bf2b201a27a40394f4a7c WatchSource:0}: Error finding container dce19ccdb807c181cb4d618ce2ead2bc30290cb5137bf2b201a27a40394f4a7c: Status 404 returned error can't find the container with id dce19ccdb807c181cb4d618ce2ead2bc30290cb5137bf2b201a27a40394f4a7c Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.277139 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:43 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:43 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:43 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.277197 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.302732 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:43 crc kubenswrapper[4726]: E1123 20:10:43.303052 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.803039864 +0000 UTC m=+151.952080820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.405761 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:43 crc kubenswrapper[4726]: E1123 20:10:43.406809 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.906773042 +0000 UTC m=+152.055814008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.407148 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:43 crc kubenswrapper[4726]: E1123 20:10:43.407548 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:43.907540715 +0000 UTC m=+152.056581671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.417602 4726 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.418550 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s5ddf"] Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.508816 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:43 crc kubenswrapper[4726]: E1123 20:10:43.509679 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:44.009657606 +0000 UTC m=+152.158698572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.535010 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5ddf" event={"ID":"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c","Type":"ContainerStarted","Data":"ae5113871448d9b797e9a4ab3a637d5d41468af95e33f42beb95239152668895"} Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.545208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sn98k" event={"ID":"2030e722-71ba-45d8-8d57-07a317e966e2","Type":"ContainerStarted","Data":"c8feee99ddb09109a25317c96940a03fd2a2d70ecb98a6b8215f65e7b4b1c01b"} Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.561359 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cn57x"] Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.566411 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4jxc" event={"ID":"f1364754-8bfa-45e8-8520-893acdb2837e","Type":"ContainerStarted","Data":"dce19ccdb807c181cb4d618ce2ead2bc30290cb5137bf2b201a27a40394f4a7c"} Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.583063 4726 generic.go:334] "Generic (PLEG): container finished" podID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerID="285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c" exitCode=0 Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.585370 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7csf" event={"ID":"15f81f72-746c-479f-b5b0-ba9ff588717a","Type":"ContainerDied","Data":"285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c"} Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.585511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7csf" event={"ID":"15f81f72-746c-479f-b5b0-ba9ff588717a","Type":"ContainerStarted","Data":"c06383edbaff76b6202a069b11b6ef87f3c81096614f643967afd33a6c88b047"} Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.588320 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.616096 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-sn98k" podStartSLOduration=12.616078083 podStartE2EDuration="12.616078083s" podCreationTimestamp="2025-11-23 20:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:43.570445062 +0000 UTC m=+151.719486018" watchObservedRunningTime="2025-11-23 20:10:43.616078083 +0000 UTC m=+151.765119039" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.616340 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:43 crc kubenswrapper[4726]: E1123 20:10:43.619146 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-23 20:10:44.119132163 +0000 UTC m=+152.268173119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vg8gw" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.717084 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:43 crc kubenswrapper[4726]: E1123 20:10:43.717792 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-23 20:10:44.217771951 +0000 UTC m=+152.366812907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.759955 4726 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-23T20:10:43.417629812Z","Handler":null,"Name":""} Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.762226 4726 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.762248 4726 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.819192 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.822248 4726 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.822292 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.843468 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.858112 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vg8gw\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.920195 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c969e63-8584-4613-85be-1d11f3014d13-secret-volume\") pod \"8c969e63-8584-4613-85be-1d11f3014d13\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.920329 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.920410 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c969e63-8584-4613-85be-1d11f3014d13-config-volume\") pod \"8c969e63-8584-4613-85be-1d11f3014d13\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.920450 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wns9\" (UniqueName: \"kubernetes.io/projected/8c969e63-8584-4613-85be-1d11f3014d13-kube-api-access-5wns9\") pod \"8c969e63-8584-4613-85be-1d11f3014d13\" (UID: \"8c969e63-8584-4613-85be-1d11f3014d13\") " Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.921418 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c969e63-8584-4613-85be-1d11f3014d13-config-volume" (OuterVolumeSpecName: "config-volume") pod "8c969e63-8584-4613-85be-1d11f3014d13" (UID: "8c969e63-8584-4613-85be-1d11f3014d13"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.927682 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c969e63-8584-4613-85be-1d11f3014d13-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8c969e63-8584-4613-85be-1d11f3014d13" (UID: "8c969e63-8584-4613-85be-1d11f3014d13"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.928400 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c969e63-8584-4613-85be-1d11f3014d13-kube-api-access-5wns9" (OuterVolumeSpecName: "kube-api-access-5wns9") pod "8c969e63-8584-4613-85be-1d11f3014d13" (UID: "8c969e63-8584-4613-85be-1d11f3014d13"). InnerVolumeSpecName "kube-api-access-5wns9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:10:43 crc kubenswrapper[4726]: I1123 20:10:43.928736 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.000692 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.022122 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c969e63-8584-4613-85be-1d11f3014d13-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.022149 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c969e63-8584-4613-85be-1d11f3014d13-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.022160 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wns9\" (UniqueName: \"kubernetes.io/projected/8c969e63-8584-4613-85be-1d11f3014d13-kube-api-access-5wns9\") on node \"crc\" DevicePath \"\"" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.190354 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nlztb"] Nov 23 20:10:44 crc kubenswrapper[4726]: E1123 20:10:44.190850 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c969e63-8584-4613-85be-1d11f3014d13" containerName="collect-profiles" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.190861 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c969e63-8584-4613-85be-1d11f3014d13" containerName="collect-profiles" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.190983 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c969e63-8584-4613-85be-1d11f3014d13" containerName="collect-profiles" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.191696 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.195612 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.206647 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nlztb"] Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.224045 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-utilities\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.224078 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-catalog-content\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.224129 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8kbr\" (UniqueName: \"kubernetes.io/projected/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-kube-api-access-r8kbr\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.262935 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.266058 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:44 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:44 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:44 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.266100 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.268580 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vg8gw"] Nov 23 20:10:44 crc kubenswrapper[4726]: W1123 20:10:44.275966 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc928dd1a_1fe0_49d9_8829_608312d1de6f.slice/crio-074a3f85e8ac95e22d1c0ab468ae8bca246a1fe3036c18c56d93dcda4c24c6e8 WatchSource:0}: Error finding container 074a3f85e8ac95e22d1c0ab468ae8bca246a1fe3036c18c56d93dcda4c24c6e8: Status 404 returned error can't find the container with id 074a3f85e8ac95e22d1c0ab468ae8bca246a1fe3036c18c56d93dcda4c24c6e8 Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.325707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-utilities\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.325744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-catalog-content\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.325794 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8kbr\" (UniqueName: \"kubernetes.io/projected/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-kube-api-access-r8kbr\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.327264 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-catalog-content\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.327595 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-utilities\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.348822 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8kbr\" (UniqueName: \"kubernetes.io/projected/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-kube-api-access-r8kbr\") pod \"redhat-marketplace-nlztb\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.399838 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.516032 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.574045 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8k58l" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.574481 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.597019 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.603167 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2mcdc"] Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.603339 4726 generic.go:334] "Generic (PLEG): container finished" podID="54adbeda-3036-4850-b945-5416bc4ed456" containerID="b19c183b1eede55450be27cafc7ad3869e85a79c8ea4ebbc72e00b6775e2e9d8" exitCode=0 Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.604108 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cn57x" event={"ID":"54adbeda-3036-4850-b945-5416bc4ed456","Type":"ContainerDied","Data":"b19c183b1eede55450be27cafc7ad3869e85a79c8ea4ebbc72e00b6775e2e9d8"} Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.604133 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cn57x" event={"ID":"54adbeda-3036-4850-b945-5416bc4ed456","Type":"ContainerStarted","Data":"6f1b21d97d40b17ac97aa38a2d4ba2d7cf3380af41f7a3c429e5227d37460fd1"} Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.604223 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.612511 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.612525 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h" event={"ID":"8c969e63-8584-4613-85be-1d11f3014d13","Type":"ContainerDied","Data":"085437d0a385d3056b90092abe36ddc8eff80b4275626387cf9afa022451f8b0"} Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.612565 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="085437d0a385d3056b90092abe36ddc8eff80b4275626387cf9afa022451f8b0" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.621447 4726 generic.go:334] "Generic (PLEG): container finished" podID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerID="62698fb53acb24f0c508f50e324e685293dae378ceb90507e33d2c60cdaead85" exitCode=0 Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.621513 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5ddf" event={"ID":"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c","Type":"ContainerDied","Data":"62698fb53acb24f0c508f50e324e685293dae378ceb90507e33d2c60cdaead85"} Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.621925 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mcdc"] Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.628959 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-catalog-content\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.629022 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjhtg\" (UniqueName: \"kubernetes.io/projected/7738a000-bd10-4bdd-9efd-3cc7b831c101-kube-api-access-sjhtg\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.629084 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-utilities\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.630817 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" event={"ID":"c928dd1a-1fe0-49d9-8829-608312d1de6f","Type":"ContainerStarted","Data":"ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306"} Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.630857 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" event={"ID":"c928dd1a-1fe0-49d9-8829-608312d1de6f","Type":"ContainerStarted","Data":"074a3f85e8ac95e22d1c0ab468ae8bca246a1fe3036c18c56d93dcda4c24c6e8"} Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.631422 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.653051 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1364754-8bfa-45e8-8520-893acdb2837e" containerID="67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375" exitCode=0 Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.654120 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4jxc" event={"ID":"f1364754-8bfa-45e8-8520-893acdb2837e","Type":"ContainerDied","Data":"67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375"} Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.732537 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjhtg\" (UniqueName: \"kubernetes.io/projected/7738a000-bd10-4bdd-9efd-3cc7b831c101-kube-api-access-sjhtg\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.732609 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-utilities\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.732667 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-catalog-content\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.733068 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-catalog-content\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.734289 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-utilities\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.761097 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" podStartSLOduration=132.761081208 podStartE2EDuration="2m12.761081208s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:44.75943752 +0000 UTC m=+152.908478486" watchObservedRunningTime="2025-11-23 20:10:44.761081208 +0000 UTC m=+152.910122164" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.770429 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjhtg\" (UniqueName: \"kubernetes.io/projected/7738a000-bd10-4bdd-9efd-3cc7b831c101-kube-api-access-sjhtg\") pod \"redhat-marketplace-2mcdc\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:44 crc kubenswrapper[4726]: I1123 20:10:44.939532 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.104317 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.104899 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.108283 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.108480 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.118900 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.135683 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nlztb"] Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.192262 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nc5vs"] Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.193975 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.199460 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.203659 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nc5vs"] Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.239454 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/613de313-7908-4f1d-9cd4-946e39ec2c10-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"613de313-7908-4f1d-9cd4-946e39ec2c10\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.239508 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/613de313-7908-4f1d-9cd4-946e39ec2c10-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"613de313-7908-4f1d-9cd4-946e39ec2c10\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.265754 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:45 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:45 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:45 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.265803 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.340815 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/613de313-7908-4f1d-9cd4-946e39ec2c10-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"613de313-7908-4f1d-9cd4-946e39ec2c10\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.340862 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw7md\" (UniqueName: \"kubernetes.io/projected/6a8b32c1-af97-4361-ab85-75d154b38a76-kube-api-access-qw7md\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.340902 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/613de313-7908-4f1d-9cd4-946e39ec2c10-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"613de313-7908-4f1d-9cd4-946e39ec2c10\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.340921 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-utilities\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.340947 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-catalog-content\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.341139 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/613de313-7908-4f1d-9cd4-946e39ec2c10-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"613de313-7908-4f1d-9cd4-946e39ec2c10\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.360992 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/613de313-7908-4f1d-9cd4-946e39ec2c10-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"613de313-7908-4f1d-9cd4-946e39ec2c10\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.443503 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw7md\" (UniqueName: \"kubernetes.io/projected/6a8b32c1-af97-4361-ab85-75d154b38a76-kube-api-access-qw7md\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.443558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-utilities\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.443589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-catalog-content\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.444218 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-catalog-content\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.444628 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-utilities\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.448694 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.454255 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mcdc"] Nov 23 20:10:45 crc kubenswrapper[4726]: W1123 20:10:45.455076 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7738a000_bd10_4bdd_9efd_3cc7b831c101.slice/crio-4ead49d331a9125260de190b48cf78874a328b4e2383254c29a5dc831190d055 WatchSource:0}: Error finding container 4ead49d331a9125260de190b48cf78874a328b4e2383254c29a5dc831190d055: Status 404 returned error can't find the container with id 4ead49d331a9125260de190b48cf78874a328b4e2383254c29a5dc831190d055 Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.459590 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw7md\" (UniqueName: \"kubernetes.io/projected/6a8b32c1-af97-4361-ab85-75d154b38a76-kube-api-access-qw7md\") pod \"redhat-operators-nc5vs\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.521780 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.592065 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q8jrp"] Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.594609 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.603432 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8jrp"] Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.675608 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerID="9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae" exitCode=0 Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.675657 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nlztb" event={"ID":"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5","Type":"ContainerDied","Data":"9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae"} Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.675683 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nlztb" event={"ID":"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5","Type":"ContainerStarted","Data":"de25986a85e84e1791604d5788a5426d8ef374a3e9f803fd4faf0fc77b5b395d"} Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.687424 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mcdc" event={"ID":"7738a000-bd10-4bdd-9efd-3cc7b831c101","Type":"ContainerStarted","Data":"4ead49d331a9125260de190b48cf78874a328b4e2383254c29a5dc831190d055"} Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.724338 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.755762 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcdlf\" (UniqueName: \"kubernetes.io/projected/0511e0ad-18eb-4add-9831-35762533ad09-kube-api-access-fcdlf\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.755803 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-utilities\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.755841 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-catalog-content\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.857776 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcdlf\" (UniqueName: \"kubernetes.io/projected/0511e0ad-18eb-4add-9831-35762533ad09-kube-api-access-fcdlf\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.858124 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-utilities\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.859022 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-utilities\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.859432 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-catalog-content\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.861100 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-catalog-content\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.900595 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcdlf\" (UniqueName: \"kubernetes.io/projected/0511e0ad-18eb-4add-9831-35762533ad09-kube-api-access-fcdlf\") pod \"redhat-operators-q8jrp\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:45 crc kubenswrapper[4726]: I1123 20:10:45.925984 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.114110 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nc5vs"] Nov 23 20:10:46 crc kubenswrapper[4726]: W1123 20:10:46.183642 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a8b32c1_af97_4361_ab85_75d154b38a76.slice/crio-cae1f031d0f579370b5886a9f24d28d159c092e1e4c624e29af9d4053e7d2e0b WatchSource:0}: Error finding container cae1f031d0f579370b5886a9f24d28d159c092e1e4c624e29af9d4053e7d2e0b: Status 404 returned error can't find the container with id cae1f031d0f579370b5886a9f24d28d159c092e1e4c624e29af9d4053e7d2e0b Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.269708 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:46 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:46 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:46 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.270069 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.563133 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8jrp"] Nov 23 20:10:46 crc kubenswrapper[4726]: W1123 20:10:46.613409 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0511e0ad_18eb_4add_9831_35762533ad09.slice/crio-4b09e2a01d6b1a1698eb0dc024c801e68d7aaa5256295873a9c658787da4aec5 WatchSource:0}: Error finding container 4b09e2a01d6b1a1698eb0dc024c801e68d7aaa5256295873a9c658787da4aec5: Status 404 returned error can't find the container with id 4b09e2a01d6b1a1698eb0dc024c801e68d7aaa5256295873a9c658787da4aec5 Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.714480 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerID="f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df" exitCode=0 Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.714555 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nc5vs" event={"ID":"6a8b32c1-af97-4361-ab85-75d154b38a76","Type":"ContainerDied","Data":"f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df"} Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.714822 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nc5vs" event={"ID":"6a8b32c1-af97-4361-ab85-75d154b38a76","Type":"ContainerStarted","Data":"cae1f031d0f579370b5886a9f24d28d159c092e1e4c624e29af9d4053e7d2e0b"} Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.725266 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"613de313-7908-4f1d-9cd4-946e39ec2c10","Type":"ContainerStarted","Data":"72b5dc90fb28a38e7bd98a78ebb0568bcca43f53fd758842bb5e32e7cc5550e7"} Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.725305 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"613de313-7908-4f1d-9cd4-946e39ec2c10","Type":"ContainerStarted","Data":"275fff20dd67b456ed2b91ce2e7e07159dc7706aceaeecc017b2fe9207f18166"} Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.744081 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8jrp" event={"ID":"0511e0ad-18eb-4add-9831-35762533ad09","Type":"ContainerStarted","Data":"4b09e2a01d6b1a1698eb0dc024c801e68d7aaa5256295873a9c658787da4aec5"} Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.756426 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.756407879 podStartE2EDuration="1.756407879s" podCreationTimestamp="2025-11-23 20:10:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:46.755646766 +0000 UTC m=+154.904687722" watchObservedRunningTime="2025-11-23 20:10:46.756407879 +0000 UTC m=+154.905448835" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.758114 4726 generic.go:334] "Generic (PLEG): container finished" podID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerID="a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d" exitCode=0 Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.760038 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mcdc" event={"ID":"7738a000-bd10-4bdd-9efd-3cc7b831c101","Type":"ContainerDied","Data":"a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d"} Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.787692 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.788413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.791739 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.791786 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.808284 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.808384 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.811109 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.909822 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.909969 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.910926 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:46 crc kubenswrapper[4726]: I1123 20:10:46.929280 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.135825 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.268861 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:47 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:47 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:47 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.268924 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.457118 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 23 20:10:47 crc kubenswrapper[4726]: W1123 20:10:47.515973 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc479eff2_7358_48a6_9c6e_5f6c53582cc1.slice/crio-e934c71869f7c3bb8299a7b4e50182ad2580bc7b9f9244b685dbc35597212f24 WatchSource:0}: Error finding container e934c71869f7c3bb8299a7b4e50182ad2580bc7b9f9244b685dbc35597212f24: Status 404 returned error can't find the container with id e934c71869f7c3bb8299a7b4e50182ad2580bc7b9f9244b685dbc35597212f24 Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.645140 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.649075 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-m5z9k" Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.766891 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c479eff2-7358-48a6-9c6e-5f6c53582cc1","Type":"ContainerStarted","Data":"e934c71869f7c3bb8299a7b4e50182ad2580bc7b9f9244b685dbc35597212f24"} Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.769129 4726 generic.go:334] "Generic (PLEG): container finished" podID="0511e0ad-18eb-4add-9831-35762533ad09" containerID="31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a" exitCode=0 Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.769184 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8jrp" event={"ID":"0511e0ad-18eb-4add-9831-35762533ad09","Type":"ContainerDied","Data":"31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a"} Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.771724 4726 generic.go:334] "Generic (PLEG): container finished" podID="613de313-7908-4f1d-9cd4-946e39ec2c10" containerID="72b5dc90fb28a38e7bd98a78ebb0568bcca43f53fd758842bb5e32e7cc5550e7" exitCode=0 Nov 23 20:10:47 crc kubenswrapper[4726]: I1123 20:10:47.772335 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"613de313-7908-4f1d-9cd4-946e39ec2c10","Type":"ContainerDied","Data":"72b5dc90fb28a38e7bd98a78ebb0568bcca43f53fd758842bb5e32e7cc5550e7"} Nov 23 20:10:48 crc kubenswrapper[4726]: I1123 20:10:48.267451 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:48 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:48 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:48 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:48 crc kubenswrapper[4726]: I1123 20:10:48.267500 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:48 crc kubenswrapper[4726]: I1123 20:10:48.819658 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c479eff2-7358-48a6-9c6e-5f6c53582cc1","Type":"ContainerStarted","Data":"a7989a77f7955c8cdf33920d15eadf01b7cd43573c1019314e377f15b2576f2d"} Nov 23 20:10:48 crc kubenswrapper[4726]: I1123 20:10:48.840085 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.840070945 podStartE2EDuration="2.840070945s" podCreationTimestamp="2025-11-23 20:10:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:10:48.836534741 +0000 UTC m=+156.985575697" watchObservedRunningTime="2025-11-23 20:10:48.840070945 +0000 UTC m=+156.989111901" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.193922 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.266907 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:49 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:49 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:49 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.266965 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.280316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/613de313-7908-4f1d-9cd4-946e39ec2c10-kubelet-dir\") pod \"613de313-7908-4f1d-9cd4-946e39ec2c10\" (UID: \"613de313-7908-4f1d-9cd4-946e39ec2c10\") " Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.280414 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/613de313-7908-4f1d-9cd4-946e39ec2c10-kube-api-access\") pod \"613de313-7908-4f1d-9cd4-946e39ec2c10\" (UID: \"613de313-7908-4f1d-9cd4-946e39ec2c10\") " Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.280442 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/613de313-7908-4f1d-9cd4-946e39ec2c10-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "613de313-7908-4f1d-9cd4-946e39ec2c10" (UID: "613de313-7908-4f1d-9cd4-946e39ec2c10"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.280695 4726 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/613de313-7908-4f1d-9cd4-946e39ec2c10-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.301706 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/613de313-7908-4f1d-9cd4-946e39ec2c10-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "613de313-7908-4f1d-9cd4-946e39ec2c10" (UID: "613de313-7908-4f1d-9cd4-946e39ec2c10"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.356653 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-579gb" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.383730 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/613de313-7908-4f1d-9cd4-946e39ec2c10-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.846323 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.846316 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"613de313-7908-4f1d-9cd4-946e39ec2c10","Type":"ContainerDied","Data":"275fff20dd67b456ed2b91ce2e7e07159dc7706aceaeecc017b2fe9207f18166"} Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.846724 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="275fff20dd67b456ed2b91ce2e7e07159dc7706aceaeecc017b2fe9207f18166" Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.864404 4726 generic.go:334] "Generic (PLEG): container finished" podID="c479eff2-7358-48a6-9c6e-5f6c53582cc1" containerID="a7989a77f7955c8cdf33920d15eadf01b7cd43573c1019314e377f15b2576f2d" exitCode=0 Nov 23 20:10:49 crc kubenswrapper[4726]: I1123 20:10:49.864456 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c479eff2-7358-48a6-9c6e-5f6c53582cc1","Type":"ContainerDied","Data":"a7989a77f7955c8cdf33920d15eadf01b7cd43573c1019314e377f15b2576f2d"} Nov 23 20:10:50 crc kubenswrapper[4726]: I1123 20:10:50.272025 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:50 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:50 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:50 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:50 crc kubenswrapper[4726]: I1123 20:10:50.272079 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.266105 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:51 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:51 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:51 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.266154 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.334905 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.417923 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kube-api-access\") pod \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\" (UID: \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\") " Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.417960 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kubelet-dir\") pod \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\" (UID: \"c479eff2-7358-48a6-9c6e-5f6c53582cc1\") " Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.418217 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c479eff2-7358-48a6-9c6e-5f6c53582cc1" (UID: "c479eff2-7358-48a6-9c6e-5f6c53582cc1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.439313 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c479eff2-7358-48a6-9c6e-5f6c53582cc1" (UID: "c479eff2-7358-48a6-9c6e-5f6c53582cc1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.519016 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.519050 4726 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c479eff2-7358-48a6-9c6e-5f6c53582cc1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.908252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c479eff2-7358-48a6-9c6e-5f6c53582cc1","Type":"ContainerDied","Data":"e934c71869f7c3bb8299a7b4e50182ad2580bc7b9f9244b685dbc35597212f24"} Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.908288 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e934c71869f7c3bb8299a7b4e50182ad2580bc7b9f9244b685dbc35597212f24" Nov 23 20:10:51 crc kubenswrapper[4726]: I1123 20:10:51.908337 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 23 20:10:52 crc kubenswrapper[4726]: I1123 20:10:52.264537 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:52 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:52 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:52 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:52 crc kubenswrapper[4726]: I1123 20:10:52.264615 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:52 crc kubenswrapper[4726]: I1123 20:10:52.864252 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-llmxt" Nov 23 20:10:52 crc kubenswrapper[4726]: I1123 20:10:52.870961 4726 patch_prober.go:28] interesting pod/console-f9d7485db-q4zwl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 23 20:10:52 crc kubenswrapper[4726]: I1123 20:10:52.871019 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-q4zwl" podUID="7445d8ee-ffb1-4610-ae2d-cb511edf1525" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 23 20:10:53 crc kubenswrapper[4726]: I1123 20:10:53.264913 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:53 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:53 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:53 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:53 crc kubenswrapper[4726]: I1123 20:10:53.265002 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:54 crc kubenswrapper[4726]: I1123 20:10:54.266250 4726 patch_prober.go:28] interesting pod/router-default-5444994796-4wfvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 23 20:10:54 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Nov 23 20:10:54 crc kubenswrapper[4726]: [+]process-running ok Nov 23 20:10:54 crc kubenswrapper[4726]: healthz check failed Nov 23 20:10:54 crc kubenswrapper[4726]: I1123 20:10:54.266560 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wfvf" podUID="a9efda09-ccd1-47d3-bdc1-b397b2e751f1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 23 20:10:54 crc kubenswrapper[4726]: I1123 20:10:54.684729 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:54 crc kubenswrapper[4726]: I1123 20:10:54.689982 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ada6b953-f533-4b33-b07e-7e80604fe4a1-metrics-certs\") pod \"network-metrics-daemon-hjmwb\" (UID: \"ada6b953-f533-4b33-b07e-7e80604fe4a1\") " pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:54 crc kubenswrapper[4726]: I1123 20:10:54.812368 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hjmwb" Nov 23 20:10:55 crc kubenswrapper[4726]: I1123 20:10:55.264908 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:10:55 crc kubenswrapper[4726]: I1123 20:10:55.266851 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-4wfvf" Nov 23 20:11:02 crc kubenswrapper[4726]: I1123 20:11:02.878844 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:11:02 crc kubenswrapper[4726]: I1123 20:11:02.887850 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:11:04 crc kubenswrapper[4726]: I1123 20:11:04.008348 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:11:09 crc kubenswrapper[4726]: I1123 20:11:09.043481 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:11:09 crc kubenswrapper[4726]: I1123 20:11:09.044181 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:11:14 crc kubenswrapper[4726]: I1123 20:11:14.586771 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hqxtv" Nov 23 20:11:18 crc kubenswrapper[4726]: E1123 20:11:18.938284 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 23 20:11:18 crc kubenswrapper[4726]: E1123 20:11:18.938757 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m82sj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cn57x_openshift-marketplace(54adbeda-3036-4850-b945-5416bc4ed456): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 20:11:18 crc kubenswrapper[4726]: E1123 20:11:18.940626 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cn57x" podUID="54adbeda-3036-4850-b945-5416bc4ed456" Nov 23 20:11:19 crc kubenswrapper[4726]: I1123 20:11:19.508608 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 23 20:11:22 crc kubenswrapper[4726]: E1123 20:11:22.251577 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cn57x" podUID="54adbeda-3036-4850-b945-5416bc4ed456" Nov 23 20:11:22 crc kubenswrapper[4726]: E1123 20:11:22.317962 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 23 20:11:22 crc kubenswrapper[4726]: E1123 20:11:22.318096 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fcdlf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-q8jrp_openshift-marketplace(0511e0ad-18eb-4add-9831-35762533ad09): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 20:11:22 crc kubenswrapper[4726]: E1123 20:11:22.319390 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-q8jrp" podUID="0511e0ad-18eb-4add-9831-35762533ad09" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.403605 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-q8jrp" podUID="0511e0ad-18eb-4add-9831-35762533ad09" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.512350 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.512498 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qw7md,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-nc5vs_openshift-marketplace(6a8b32c1-af97-4361-ab85-75d154b38a76): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.514442 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-nc5vs" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.517830 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.517988 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r8kbr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-nlztb_openshift-marketplace(a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.519209 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-nlztb" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.522222 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.522348 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sjhtg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2mcdc_openshift-marketplace(7738a000-bd10-4bdd-9efd-3cc7b831c101): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.523499 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2mcdc" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.764545 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2mcdc" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.764604 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-nlztb" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" Nov 23 20:11:23 crc kubenswrapper[4726]: E1123 20:11:23.773441 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-nc5vs" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" Nov 23 20:11:23 crc kubenswrapper[4726]: I1123 20:11:23.958514 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hjmwb"] Nov 23 20:11:23 crc kubenswrapper[4726]: W1123 20:11:23.965379 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podada6b953_f533_4b33_b07e_7e80604fe4a1.slice/crio-c6bd7d43833ed14e8f967b7d3e33d2e77a41f2bf557366e1dc6640448b04af97 WatchSource:0}: Error finding container c6bd7d43833ed14e8f967b7d3e33d2e77a41f2bf557366e1dc6640448b04af97: Status 404 returned error can't find the container with id c6bd7d43833ed14e8f967b7d3e33d2e77a41f2bf557366e1dc6640448b04af97 Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.768032 4726 generic.go:334] "Generic (PLEG): container finished" podID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerID="692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212" exitCode=0 Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.768517 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7csf" event={"ID":"15f81f72-746c-479f-b5b0-ba9ff588717a","Type":"ContainerDied","Data":"692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212"} Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.772630 4726 generic.go:334] "Generic (PLEG): container finished" podID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerID="59943389ee490d3f53f3069856b2f982089b844302374a17bbc61d81b1fe9842" exitCode=0 Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.772710 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5ddf" event={"ID":"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c","Type":"ContainerDied","Data":"59943389ee490d3f53f3069856b2f982089b844302374a17bbc61d81b1fe9842"} Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.774924 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" event={"ID":"ada6b953-f533-4b33-b07e-7e80604fe4a1","Type":"ContainerStarted","Data":"e1d3e5afd6b87235ef6211baf52a975811ffbf82ae9f3b319bfc77f81a1d440b"} Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.774964 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" event={"ID":"ada6b953-f533-4b33-b07e-7e80604fe4a1","Type":"ContainerStarted","Data":"01d01ea863dc669a94236ae32e3ae41a50fa116cce45e412d1d75565fd6c6b2c"} Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.774981 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hjmwb" event={"ID":"ada6b953-f533-4b33-b07e-7e80604fe4a1","Type":"ContainerStarted","Data":"c6bd7d43833ed14e8f967b7d3e33d2e77a41f2bf557366e1dc6640448b04af97"} Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.780301 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1364754-8bfa-45e8-8520-893acdb2837e" containerID="29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0" exitCode=0 Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.780340 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4jxc" event={"ID":"f1364754-8bfa-45e8-8520-893acdb2837e","Type":"ContainerDied","Data":"29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0"} Nov 23 20:11:24 crc kubenswrapper[4726]: I1123 20:11:24.850936 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hjmwb" podStartSLOduration=172.850919594 podStartE2EDuration="2m52.850919594s" podCreationTimestamp="2025-11-23 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:11:24.849181773 +0000 UTC m=+192.998222739" watchObservedRunningTime="2025-11-23 20:11:24.850919594 +0000 UTC m=+192.999960550" Nov 23 20:11:25 crc kubenswrapper[4726]: I1123 20:11:25.786954 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5ddf" event={"ID":"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c","Type":"ContainerStarted","Data":"fd4f66fb85243486ed022a3c3c3d8c6c5c071523f7f7e7c6d798e1464b165c6b"} Nov 23 20:11:25 crc kubenswrapper[4726]: I1123 20:11:25.789804 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4jxc" event={"ID":"f1364754-8bfa-45e8-8520-893acdb2837e","Type":"ContainerStarted","Data":"78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45"} Nov 23 20:11:25 crc kubenswrapper[4726]: I1123 20:11:25.791802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7csf" event={"ID":"15f81f72-746c-479f-b5b0-ba9ff588717a","Type":"ContainerStarted","Data":"28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c"} Nov 23 20:11:25 crc kubenswrapper[4726]: I1123 20:11:25.839958 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s5ddf" podStartSLOduration=3.233106138 podStartE2EDuration="43.839940295s" podCreationTimestamp="2025-11-23 20:10:42 +0000 UTC" firstStartedPulling="2025-11-23 20:10:44.628234625 +0000 UTC m=+152.777275581" lastFinishedPulling="2025-11-23 20:11:25.235068782 +0000 UTC m=+193.384109738" observedRunningTime="2025-11-23 20:11:25.822086171 +0000 UTC m=+193.971127127" watchObservedRunningTime="2025-11-23 20:11:25.839940295 +0000 UTC m=+193.988981251" Nov 23 20:11:25 crc kubenswrapper[4726]: I1123 20:11:25.841419 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h4jxc" podStartSLOduration=3.246599195 podStartE2EDuration="43.841413719s" podCreationTimestamp="2025-11-23 20:10:42 +0000 UTC" firstStartedPulling="2025-11-23 20:10:44.656839815 +0000 UTC m=+152.805880771" lastFinishedPulling="2025-11-23 20:11:25.251654339 +0000 UTC m=+193.400695295" observedRunningTime="2025-11-23 20:11:25.84044372 +0000 UTC m=+193.989484676" watchObservedRunningTime="2025-11-23 20:11:25.841413719 +0000 UTC m=+193.990454685" Nov 23 20:11:25 crc kubenswrapper[4726]: I1123 20:11:25.859537 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t7csf" podStartSLOduration=3.283604668 podStartE2EDuration="44.85950714s" podCreationTimestamp="2025-11-23 20:10:41 +0000 UTC" firstStartedPulling="2025-11-23 20:10:43.588027049 +0000 UTC m=+151.737068005" lastFinishedPulling="2025-11-23 20:11:25.163929521 +0000 UTC m=+193.312970477" observedRunningTime="2025-11-23 20:11:25.857709608 +0000 UTC m=+194.006750564" watchObservedRunningTime="2025-11-23 20:11:25.85950714 +0000 UTC m=+194.008548096" Nov 23 20:11:27 crc kubenswrapper[4726]: I1123 20:11:27.164792 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vmvvq"] Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.309707 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.310056 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.479888 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.536878 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.536911 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.574044 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.718618 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.718664 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.752104 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.880462 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.883272 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:11:32 crc kubenswrapper[4726]: I1123 20:11:32.890933 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:11:34 crc kubenswrapper[4726]: I1123 20:11:34.508818 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s5ddf"] Nov 23 20:11:34 crc kubenswrapper[4726]: I1123 20:11:34.856661 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s5ddf" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerName="registry-server" containerID="cri-o://fd4f66fb85243486ed022a3c3c3d8c6c5c071523f7f7e7c6d798e1464b165c6b" gracePeriod=2 Nov 23 20:11:35 crc kubenswrapper[4726]: I1123 20:11:35.864171 4726 generic.go:334] "Generic (PLEG): container finished" podID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerID="fd4f66fb85243486ed022a3c3c3d8c6c5c071523f7f7e7c6d798e1464b165c6b" exitCode=0 Nov 23 20:11:35 crc kubenswrapper[4726]: I1123 20:11:35.864230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5ddf" event={"ID":"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c","Type":"ContainerDied","Data":"fd4f66fb85243486ed022a3c3c3d8c6c5c071523f7f7e7c6d798e1464b165c6b"} Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.523452 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.550762 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-catalog-content\") pod \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.550845 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-utilities\") pod \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.550905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs7gr\" (UniqueName: \"kubernetes.io/projected/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-kube-api-access-vs7gr\") pod \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\" (UID: \"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c\") " Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.552285 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-utilities" (OuterVolumeSpecName: "utilities") pod "3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" (UID: "3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.571347 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-kube-api-access-vs7gr" (OuterVolumeSpecName: "kube-api-access-vs7gr") pod "3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" (UID: "3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c"). InnerVolumeSpecName "kube-api-access-vs7gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.624335 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" (UID: "3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.652000 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.652037 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs7gr\" (UniqueName: \"kubernetes.io/projected/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-kube-api-access-vs7gr\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.652053 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.870736 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nc5vs" event={"ID":"6a8b32c1-af97-4361-ab85-75d154b38a76","Type":"ContainerStarted","Data":"d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e"} Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.874039 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5ddf" event={"ID":"3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c","Type":"ContainerDied","Data":"ae5113871448d9b797e9a4ab3a637d5d41468af95e33f42beb95239152668895"} Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.874070 4726 scope.go:117] "RemoveContainer" containerID="fd4f66fb85243486ed022a3c3c3d8c6c5c071523f7f7e7c6d798e1464b165c6b" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.874106 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s5ddf" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.913991 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s5ddf"] Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.916440 4726 scope.go:117] "RemoveContainer" containerID="59943389ee490d3f53f3069856b2f982089b844302374a17bbc61d81b1fe9842" Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.917324 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s5ddf"] Nov 23 20:11:36 crc kubenswrapper[4726]: I1123 20:11:36.959832 4726 scope.go:117] "RemoveContainer" containerID="62698fb53acb24f0c508f50e324e685293dae378ceb90507e33d2c60cdaead85" Nov 23 20:11:37 crc kubenswrapper[4726]: I1123 20:11:37.881913 4726 generic.go:334] "Generic (PLEG): container finished" podID="54adbeda-3036-4850-b945-5416bc4ed456" containerID="6eb07fe531cebe752f9ffccdb0d64a70ba3ca9108d70927b03b1be4565f115c6" exitCode=0 Nov 23 20:11:37 crc kubenswrapper[4726]: I1123 20:11:37.881996 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cn57x" event={"ID":"54adbeda-3036-4850-b945-5416bc4ed456","Type":"ContainerDied","Data":"6eb07fe531cebe752f9ffccdb0d64a70ba3ca9108d70927b03b1be4565f115c6"} Nov 23 20:11:37 crc kubenswrapper[4726]: I1123 20:11:37.885820 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerID="d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e" exitCode=0 Nov 23 20:11:37 crc kubenswrapper[4726]: I1123 20:11:37.885906 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nc5vs" event={"ID":"6a8b32c1-af97-4361-ab85-75d154b38a76","Type":"ContainerDied","Data":"d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e"} Nov 23 20:11:37 crc kubenswrapper[4726]: I1123 20:11:37.892520 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerID="44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949" exitCode=0 Nov 23 20:11:37 crc kubenswrapper[4726]: I1123 20:11:37.892557 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nlztb" event={"ID":"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5","Type":"ContainerDied","Data":"44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949"} Nov 23 20:11:37 crc kubenswrapper[4726]: I1123 20:11:37.900677 4726 generic.go:334] "Generic (PLEG): container finished" podID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerID="e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee" exitCode=0 Nov 23 20:11:37 crc kubenswrapper[4726]: I1123 20:11:37.900707 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mcdc" event={"ID":"7738a000-bd10-4bdd-9efd-3cc7b831c101","Type":"ContainerDied","Data":"e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee"} Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.596214 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" path="/var/lib/kubelet/pods/3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c/volumes" Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.907045 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mcdc" event={"ID":"7738a000-bd10-4bdd-9efd-3cc7b831c101","Type":"ContainerStarted","Data":"d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c"} Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.910036 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cn57x" event={"ID":"54adbeda-3036-4850-b945-5416bc4ed456","Type":"ContainerStarted","Data":"87d29d2cc1ac1ef1eb34a068175e74be9814c7e0ad0f95d1525882eb18104aaf"} Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.912663 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nc5vs" event={"ID":"6a8b32c1-af97-4361-ab85-75d154b38a76","Type":"ContainerStarted","Data":"7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9"} Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.914626 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nlztb" event={"ID":"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5","Type":"ContainerStarted","Data":"da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712"} Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.924811 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2mcdc" podStartSLOduration=3.326861421 podStartE2EDuration="54.924797496s" podCreationTimestamp="2025-11-23 20:10:44 +0000 UTC" firstStartedPulling="2025-11-23 20:10:46.764518377 +0000 UTC m=+154.913559333" lastFinishedPulling="2025-11-23 20:11:38.362454452 +0000 UTC m=+206.511495408" observedRunningTime="2025-11-23 20:11:38.922479367 +0000 UTC m=+207.071520323" watchObservedRunningTime="2025-11-23 20:11:38.924797496 +0000 UTC m=+207.073838452" Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.953211 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cn57x" podStartSLOduration=3.240182847 podStartE2EDuration="56.953196313s" podCreationTimestamp="2025-11-23 20:10:42 +0000 UTC" firstStartedPulling="2025-11-23 20:10:44.605046953 +0000 UTC m=+152.754087909" lastFinishedPulling="2025-11-23 20:11:38.318060419 +0000 UTC m=+206.467101375" observedRunningTime="2025-11-23 20:11:38.948588685 +0000 UTC m=+207.097629641" watchObservedRunningTime="2025-11-23 20:11:38.953196313 +0000 UTC m=+207.102237269" Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.990766 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nlztb" podStartSLOduration=2.394481376 podStartE2EDuration="54.990748861s" podCreationTimestamp="2025-11-23 20:10:44 +0000 UTC" firstStartedPulling="2025-11-23 20:10:45.678095263 +0000 UTC m=+153.827136219" lastFinishedPulling="2025-11-23 20:11:38.274362748 +0000 UTC m=+206.423403704" observedRunningTime="2025-11-23 20:11:38.973821467 +0000 UTC m=+207.122862433" watchObservedRunningTime="2025-11-23 20:11:38.990748861 +0000 UTC m=+207.139789818" Nov 23 20:11:38 crc kubenswrapper[4726]: I1123 20:11:38.992751 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nc5vs" podStartSLOduration=2.36341337 podStartE2EDuration="53.992745341s" podCreationTimestamp="2025-11-23 20:10:45 +0000 UTC" firstStartedPulling="2025-11-23 20:10:46.717166286 +0000 UTC m=+154.866207242" lastFinishedPulling="2025-11-23 20:11:38.346498267 +0000 UTC m=+206.495539213" observedRunningTime="2025-11-23 20:11:38.989210826 +0000 UTC m=+207.138251782" watchObservedRunningTime="2025-11-23 20:11:38.992745341 +0000 UTC m=+207.141786287" Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.043539 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.043613 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.043674 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.044307 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.044430 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60" gracePeriod=600 Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.920503 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8jrp" event={"ID":"0511e0ad-18eb-4add-9831-35762533ad09","Type":"ContainerStarted","Data":"8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f"} Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.923664 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60" exitCode=0 Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.923710 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60"} Nov 23 20:11:39 crc kubenswrapper[4726]: I1123 20:11:39.923736 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"d69e63c1f53e98006b9aa5285330cc02cef0dd0c90e397e35c78c1a1ac967d6f"} Nov 23 20:11:40 crc kubenswrapper[4726]: I1123 20:11:40.931053 4726 generic.go:334] "Generic (PLEG): container finished" podID="0511e0ad-18eb-4add-9831-35762533ad09" containerID="8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f" exitCode=0 Nov 23 20:11:40 crc kubenswrapper[4726]: I1123 20:11:40.931094 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8jrp" event={"ID":"0511e0ad-18eb-4add-9831-35762533ad09","Type":"ContainerDied","Data":"8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f"} Nov 23 20:11:41 crc kubenswrapper[4726]: I1123 20:11:41.937633 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8jrp" event={"ID":"0511e0ad-18eb-4add-9831-35762533ad09","Type":"ContainerStarted","Data":"7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0"} Nov 23 20:11:41 crc kubenswrapper[4726]: I1123 20:11:41.962357 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q8jrp" podStartSLOduration=3.367340377 podStartE2EDuration="56.962339051s" podCreationTimestamp="2025-11-23 20:10:45 +0000 UTC" firstStartedPulling="2025-11-23 20:10:47.776969467 +0000 UTC m=+155.926010423" lastFinishedPulling="2025-11-23 20:11:41.371968141 +0000 UTC m=+209.521009097" observedRunningTime="2025-11-23 20:11:41.959224088 +0000 UTC m=+210.108265054" watchObservedRunningTime="2025-11-23 20:11:41.962339051 +0000 UTC m=+210.111380017" Nov 23 20:11:42 crc kubenswrapper[4726]: I1123 20:11:42.944370 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:11:42 crc kubenswrapper[4726]: I1123 20:11:42.944423 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:11:42 crc kubenswrapper[4726]: I1123 20:11:42.984384 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:11:43 crc kubenswrapper[4726]: I1123 20:11:43.992540 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:11:44 crc kubenswrapper[4726]: I1123 20:11:44.518296 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:11:44 crc kubenswrapper[4726]: I1123 20:11:44.518594 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:11:44 crc kubenswrapper[4726]: I1123 20:11:44.571468 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:11:44 crc kubenswrapper[4726]: I1123 20:11:44.940016 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:11:44 crc kubenswrapper[4726]: I1123 20:11:44.940302 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:11:44 crc kubenswrapper[4726]: I1123 20:11:44.997843 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:11:45 crc kubenswrapper[4726]: I1123 20:11:45.010556 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:11:45 crc kubenswrapper[4726]: I1123 20:11:45.523268 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:11:45 crc kubenswrapper[4726]: I1123 20:11:45.524108 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:11:45 crc kubenswrapper[4726]: I1123 20:11:45.583639 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:11:45 crc kubenswrapper[4726]: I1123 20:11:45.927058 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:11:45 crc kubenswrapper[4726]: I1123 20:11:45.927124 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:11:46 crc kubenswrapper[4726]: I1123 20:11:46.018475 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:11:46 crc kubenswrapper[4726]: I1123 20:11:46.018850 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:11:46 crc kubenswrapper[4726]: I1123 20:11:46.989814 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q8jrp" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="registry-server" probeResult="failure" output=< Nov 23 20:11:46 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 20:11:46 crc kubenswrapper[4726]: > Nov 23 20:11:47 crc kubenswrapper[4726]: I1123 20:11:47.114332 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cn57x"] Nov 23 20:11:47 crc kubenswrapper[4726]: I1123 20:11:47.114689 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cn57x" podUID="54adbeda-3036-4850-b945-5416bc4ed456" containerName="registry-server" containerID="cri-o://87d29d2cc1ac1ef1eb34a068175e74be9814c7e0ad0f95d1525882eb18104aaf" gracePeriod=2 Nov 23 20:11:47 crc kubenswrapper[4726]: I1123 20:11:47.310002 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mcdc"] Nov 23 20:11:47 crc kubenswrapper[4726]: I1123 20:11:47.980139 4726 generic.go:334] "Generic (PLEG): container finished" podID="54adbeda-3036-4850-b945-5416bc4ed456" containerID="87d29d2cc1ac1ef1eb34a068175e74be9814c7e0ad0f95d1525882eb18104aaf" exitCode=0 Nov 23 20:11:47 crc kubenswrapper[4726]: I1123 20:11:47.980248 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cn57x" event={"ID":"54adbeda-3036-4850-b945-5416bc4ed456","Type":"ContainerDied","Data":"87d29d2cc1ac1ef1eb34a068175e74be9814c7e0ad0f95d1525882eb18104aaf"} Nov 23 20:11:47 crc kubenswrapper[4726]: I1123 20:11:47.980369 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2mcdc" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerName="registry-server" containerID="cri-o://d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c" gracePeriod=2 Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.255047 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.363078 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.448213 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-catalog-content\") pod \"54adbeda-3036-4850-b945-5416bc4ed456\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.448302 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-utilities\") pod \"54adbeda-3036-4850-b945-5416bc4ed456\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.448333 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m82sj\" (UniqueName: \"kubernetes.io/projected/54adbeda-3036-4850-b945-5416bc4ed456-kube-api-access-m82sj\") pod \"54adbeda-3036-4850-b945-5416bc4ed456\" (UID: \"54adbeda-3036-4850-b945-5416bc4ed456\") " Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.448558 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-utilities\") pod \"7738a000-bd10-4bdd-9efd-3cc7b831c101\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.448622 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-catalog-content\") pod \"7738a000-bd10-4bdd-9efd-3cc7b831c101\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.449403 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-utilities" (OuterVolumeSpecName: "utilities") pod "7738a000-bd10-4bdd-9efd-3cc7b831c101" (UID: "7738a000-bd10-4bdd-9efd-3cc7b831c101"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.449505 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-utilities" (OuterVolumeSpecName: "utilities") pod "54adbeda-3036-4850-b945-5416bc4ed456" (UID: "54adbeda-3036-4850-b945-5416bc4ed456"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.453933 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54adbeda-3036-4850-b945-5416bc4ed456-kube-api-access-m82sj" (OuterVolumeSpecName: "kube-api-access-m82sj") pod "54adbeda-3036-4850-b945-5416bc4ed456" (UID: "54adbeda-3036-4850-b945-5416bc4ed456"). InnerVolumeSpecName "kube-api-access-m82sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.466617 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7738a000-bd10-4bdd-9efd-3cc7b831c101" (UID: "7738a000-bd10-4bdd-9efd-3cc7b831c101"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.498368 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54adbeda-3036-4850-b945-5416bc4ed456" (UID: "54adbeda-3036-4850-b945-5416bc4ed456"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.550701 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjhtg\" (UniqueName: \"kubernetes.io/projected/7738a000-bd10-4bdd-9efd-3cc7b831c101-kube-api-access-sjhtg\") pod \"7738a000-bd10-4bdd-9efd-3cc7b831c101\" (UID: \"7738a000-bd10-4bdd-9efd-3cc7b831c101\") " Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.551440 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.551508 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.551530 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54adbeda-3036-4850-b945-5416bc4ed456-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.551550 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m82sj\" (UniqueName: \"kubernetes.io/projected/54adbeda-3036-4850-b945-5416bc4ed456-kube-api-access-m82sj\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.551605 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7738a000-bd10-4bdd-9efd-3cc7b831c101-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.553403 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7738a000-bd10-4bdd-9efd-3cc7b831c101-kube-api-access-sjhtg" (OuterVolumeSpecName: "kube-api-access-sjhtg") pod "7738a000-bd10-4bdd-9efd-3cc7b831c101" (UID: "7738a000-bd10-4bdd-9efd-3cc7b831c101"). InnerVolumeSpecName "kube-api-access-sjhtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.652966 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjhtg\" (UniqueName: \"kubernetes.io/projected/7738a000-bd10-4bdd-9efd-3cc7b831c101-kube-api-access-sjhtg\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.990401 4726 generic.go:334] "Generic (PLEG): container finished" podID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerID="d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c" exitCode=0 Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.990466 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mcdc" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.990494 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mcdc" event={"ID":"7738a000-bd10-4bdd-9efd-3cc7b831c101","Type":"ContainerDied","Data":"d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c"} Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.990566 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mcdc" event={"ID":"7738a000-bd10-4bdd-9efd-3cc7b831c101","Type":"ContainerDied","Data":"4ead49d331a9125260de190b48cf78874a328b4e2383254c29a5dc831190d055"} Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.990611 4726 scope.go:117] "RemoveContainer" containerID="d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c" Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.997366 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cn57x" event={"ID":"54adbeda-3036-4850-b945-5416bc4ed456","Type":"ContainerDied","Data":"6f1b21d97d40b17ac97aa38a2d4ba2d7cf3380af41f7a3c429e5227d37460fd1"} Nov 23 20:11:48 crc kubenswrapper[4726]: I1123 20:11:48.997447 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cn57x" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.027061 4726 scope.go:117] "RemoveContainer" containerID="e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.043397 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mcdc"] Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.057522 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mcdc"] Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.057634 4726 scope.go:117] "RemoveContainer" containerID="a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.063924 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cn57x"] Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.068782 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cn57x"] Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.078545 4726 scope.go:117] "RemoveContainer" containerID="d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c" Nov 23 20:11:49 crc kubenswrapper[4726]: E1123 20:11:49.079183 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c\": container with ID starting with d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c not found: ID does not exist" containerID="d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.079234 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c"} err="failed to get container status \"d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c\": rpc error: code = NotFound desc = could not find container \"d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c\": container with ID starting with d722fee24cdedc9430021175d735bb8fc6847c41e2ed8fe9759899d6fbf8818c not found: ID does not exist" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.079268 4726 scope.go:117] "RemoveContainer" containerID="e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee" Nov 23 20:11:49 crc kubenswrapper[4726]: E1123 20:11:49.079552 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee\": container with ID starting with e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee not found: ID does not exist" containerID="e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.079583 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee"} err="failed to get container status \"e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee\": rpc error: code = NotFound desc = could not find container \"e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee\": container with ID starting with e0999ed9b1185faf1aad3e7335730ea81b6e612ea57e65fdc3dff08bdf0a55ee not found: ID does not exist" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.079610 4726 scope.go:117] "RemoveContainer" containerID="a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d" Nov 23 20:11:49 crc kubenswrapper[4726]: E1123 20:11:49.079972 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d\": container with ID starting with a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d not found: ID does not exist" containerID="a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.080008 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d"} err="failed to get container status \"a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d\": rpc error: code = NotFound desc = could not find container \"a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d\": container with ID starting with a40fed815e6e5567523e86a8a8516ad2f0904aebd8234ed49cae49ac2431355d not found: ID does not exist" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.080031 4726 scope.go:117] "RemoveContainer" containerID="87d29d2cc1ac1ef1eb34a068175e74be9814c7e0ad0f95d1525882eb18104aaf" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.106069 4726 scope.go:117] "RemoveContainer" containerID="6eb07fe531cebe752f9ffccdb0d64a70ba3ca9108d70927b03b1be4565f115c6" Nov 23 20:11:49 crc kubenswrapper[4726]: I1123 20:11:49.124975 4726 scope.go:117] "RemoveContainer" containerID="b19c183b1eede55450be27cafc7ad3869e85a79c8ea4ebbc72e00b6775e2e9d8" Nov 23 20:11:50 crc kubenswrapper[4726]: I1123 20:11:50.599708 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54adbeda-3036-4850-b945-5416bc4ed456" path="/var/lib/kubelet/pods/54adbeda-3036-4850-b945-5416bc4ed456/volumes" Nov 23 20:11:50 crc kubenswrapper[4726]: I1123 20:11:50.602561 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" path="/var/lib/kubelet/pods/7738a000-bd10-4bdd-9efd-3cc7b831c101/volumes" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.199662 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" podUID="187fba44-db16-4325-9df5-75ab2f9da2d3" containerName="oauth-openshift" containerID="cri-o://c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660" gracePeriod=15 Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.673692 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712281 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-error\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712349 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-dir\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712387 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-session\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712442 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqzqn\" (UniqueName: \"kubernetes.io/projected/187fba44-db16-4325-9df5-75ab2f9da2d3-kube-api-access-fqzqn\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712472 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-serving-cert\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712505 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-trusted-ca-bundle\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712504 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712538 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-router-certs\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712565 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-login\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712615 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-idp-0-file-data\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712648 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-ocp-branding-template\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712671 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-policies\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712703 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-provider-selection\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712735 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-cliconfig\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.712767 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-service-ca\") pod \"187fba44-db16-4325-9df5-75ab2f9da2d3\" (UID: \"187fba44-db16-4325-9df5-75ab2f9da2d3\") " Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.713035 4726 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.713685 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.716230 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.718474 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.718875 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.722149 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.722496 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.725598 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.726118 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/187fba44-db16-4325-9df5-75ab2f9da2d3-kube-api-access-fqzqn" (OuterVolumeSpecName: "kube-api-access-fqzqn") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "kube-api-access-fqzqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.730241 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.730376 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.731607 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.732122 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.737194 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "187fba44-db16-4325-9df5-75ab2f9da2d3" (UID: "187fba44-db16-4325-9df5-75ab2f9da2d3"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814329 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814410 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814429 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814443 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814483 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814496 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814507 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814518 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqzqn\" (UniqueName: \"kubernetes.io/projected/187fba44-db16-4325-9df5-75ab2f9da2d3-kube-api-access-fqzqn\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814529 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814568 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814580 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814591 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.814603 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/187fba44-db16-4325-9df5-75ab2f9da2d3-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.878914 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-kn7pz"] Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879104 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54adbeda-3036-4850-b945-5416bc4ed456" containerName="extract-content" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879116 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="54adbeda-3036-4850-b945-5416bc4ed456" containerName="extract-content" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879126 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerName="extract-utilities" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879131 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerName="extract-utilities" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879141 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="187fba44-db16-4325-9df5-75ab2f9da2d3" containerName="oauth-openshift" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879147 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="187fba44-db16-4325-9df5-75ab2f9da2d3" containerName="oauth-openshift" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879156 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerName="extract-content" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879162 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerName="extract-content" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879170 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="613de313-7908-4f1d-9cd4-946e39ec2c10" containerName="pruner" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879176 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="613de313-7908-4f1d-9cd4-946e39ec2c10" containerName="pruner" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879183 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879189 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879202 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerName="extract-content" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879207 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerName="extract-content" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879215 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerName="extract-utilities" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879223 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerName="extract-utilities" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879232 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c479eff2-7358-48a6-9c6e-5f6c53582cc1" containerName="pruner" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879238 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c479eff2-7358-48a6-9c6e-5f6c53582cc1" containerName="pruner" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879245 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54adbeda-3036-4850-b945-5416bc4ed456" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879250 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="54adbeda-3036-4850-b945-5416bc4ed456" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879257 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879262 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: E1123 20:11:52.879269 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54adbeda-3036-4850-b945-5416bc4ed456" containerName="extract-utilities" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879275 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="54adbeda-3036-4850-b945-5416bc4ed456" containerName="extract-utilities" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879355 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="613de313-7908-4f1d-9cd4-946e39ec2c10" containerName="pruner" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879368 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c479eff2-7358-48a6-9c6e-5f6c53582cc1" containerName="pruner" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879376 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="54adbeda-3036-4850-b945-5416bc4ed456" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879383 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b33f1f1-1b91-4706-bf1d-a90d0cb0a88c" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879391 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7738a000-bd10-4bdd-9efd-3cc7b831c101" containerName="registry-server" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879398 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="187fba44-db16-4325-9df5-75ab2f9da2d3" containerName="oauth-openshift" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.879779 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.888546 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-kn7pz"] Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.915711 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.915848 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.915933 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916039 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916085 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916216 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916275 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916337 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916387 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916409 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916441 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-audit-policies\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916460 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzwjh\" (UniqueName: \"kubernetes.io/projected/0da86596-1cde-4389-aec7-46b8ecc136f0-kube-api-access-lzwjh\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:52 crc kubenswrapper[4726]: I1123 20:11:52.916488 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0da86596-1cde-4389-aec7-46b8ecc136f0-audit-dir\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017419 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017483 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017530 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017616 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017636 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzwjh\" (UniqueName: \"kubernetes.io/projected/0da86596-1cde-4389-aec7-46b8ecc136f0-kube-api-access-lzwjh\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017657 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-audit-policies\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017682 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0da86596-1cde-4389-aec7-46b8ecc136f0-audit-dir\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017716 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017772 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.017839 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.018123 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0da86596-1cde-4389-aec7-46b8ecc136f0-audit-dir\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.019429 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-audit-policies\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.020332 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.019450 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.021045 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.021757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.021845 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.023171 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.023475 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.023760 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.026143 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.028357 4726 generic.go:334] "Generic (PLEG): container finished" podID="187fba44-db16-4325-9df5-75ab2f9da2d3" containerID="c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660" exitCode=0 Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.028397 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" event={"ID":"187fba44-db16-4325-9df5-75ab2f9da2d3","Type":"ContainerDied","Data":"c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660"} Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.028428 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" event={"ID":"187fba44-db16-4325-9df5-75ab2f9da2d3","Type":"ContainerDied","Data":"8a50c562fd384968d05d5fcf20f92b1eb5b5dbcf3eb1866a2ab70537912c0b24"} Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.028449 4726 scope.go:117] "RemoveContainer" containerID="c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.028570 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vmvvq" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.028616 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.031447 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0da86596-1cde-4389-aec7-46b8ecc136f0-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.038088 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzwjh\" (UniqueName: \"kubernetes.io/projected/0da86596-1cde-4389-aec7-46b8ecc136f0-kube-api-access-lzwjh\") pod \"oauth-openshift-785f95f6b7-kn7pz\" (UID: \"0da86596-1cde-4389-aec7-46b8ecc136f0\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.053353 4726 scope.go:117] "RemoveContainer" containerID="c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660" Nov 23 20:11:53 crc kubenswrapper[4726]: E1123 20:11:53.054196 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660\": container with ID starting with c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660 not found: ID does not exist" containerID="c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.054242 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660"} err="failed to get container status \"c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660\": rpc error: code = NotFound desc = could not find container \"c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660\": container with ID starting with c92ed9d27aa284a515b3f241a04c1bc6a30465e5212a7f346a5be3b98b7e5660 not found: ID does not exist" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.071963 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vmvvq"] Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.079800 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vmvvq"] Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.210802 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:53 crc kubenswrapper[4726]: I1123 20:11:53.615324 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-kn7pz"] Nov 23 20:11:53 crc kubenswrapper[4726]: W1123 20:11:53.622228 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0da86596_1cde_4389_aec7_46b8ecc136f0.slice/crio-94132204fb2ecbca8441dfbacdf804d77b795bb0791cb7f595ed0a9ccba06192 WatchSource:0}: Error finding container 94132204fb2ecbca8441dfbacdf804d77b795bb0791cb7f595ed0a9ccba06192: Status 404 returned error can't find the container with id 94132204fb2ecbca8441dfbacdf804d77b795bb0791cb7f595ed0a9ccba06192 Nov 23 20:11:54 crc kubenswrapper[4726]: I1123 20:11:54.033462 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" event={"ID":"0da86596-1cde-4389-aec7-46b8ecc136f0","Type":"ContainerStarted","Data":"94132204fb2ecbca8441dfbacdf804d77b795bb0791cb7f595ed0a9ccba06192"} Nov 23 20:11:54 crc kubenswrapper[4726]: I1123 20:11:54.594893 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="187fba44-db16-4325-9df5-75ab2f9da2d3" path="/var/lib/kubelet/pods/187fba44-db16-4325-9df5-75ab2f9da2d3/volumes" Nov 23 20:11:55 crc kubenswrapper[4726]: I1123 20:11:55.040161 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" event={"ID":"0da86596-1cde-4389-aec7-46b8ecc136f0","Type":"ContainerStarted","Data":"0be4dc1b00da83979b54b448b5a1593b4fe31d52261ed99d3e7cd4d34f690f0e"} Nov 23 20:11:55 crc kubenswrapper[4726]: I1123 20:11:55.040368 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:55 crc kubenswrapper[4726]: I1123 20:11:55.048445 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" Nov 23 20:11:55 crc kubenswrapper[4726]: I1123 20:11:55.067966 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-785f95f6b7-kn7pz" podStartSLOduration=28.067949142 podStartE2EDuration="28.067949142s" podCreationTimestamp="2025-11-23 20:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:11:55.064423397 +0000 UTC m=+223.213464363" watchObservedRunningTime="2025-11-23 20:11:55.067949142 +0000 UTC m=+223.216990098" Nov 23 20:11:55 crc kubenswrapper[4726]: I1123 20:11:55.959730 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:11:55 crc kubenswrapper[4726]: I1123 20:11:55.994405 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:11:58 crc kubenswrapper[4726]: I1123 20:11:58.360766 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8jrp"] Nov 23 20:11:58 crc kubenswrapper[4726]: I1123 20:11:58.361567 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q8jrp" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="registry-server" containerID="cri-o://7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0" gracePeriod=2 Nov 23 20:11:58 crc kubenswrapper[4726]: I1123 20:11:58.794441 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:11:58 crc kubenswrapper[4726]: I1123 20:11:58.904539 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-utilities\") pod \"0511e0ad-18eb-4add-9831-35762533ad09\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " Nov 23 20:11:58 crc kubenswrapper[4726]: I1123 20:11:58.904626 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-catalog-content\") pod \"0511e0ad-18eb-4add-9831-35762533ad09\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " Nov 23 20:11:58 crc kubenswrapper[4726]: I1123 20:11:58.904666 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcdlf\" (UniqueName: \"kubernetes.io/projected/0511e0ad-18eb-4add-9831-35762533ad09-kube-api-access-fcdlf\") pod \"0511e0ad-18eb-4add-9831-35762533ad09\" (UID: \"0511e0ad-18eb-4add-9831-35762533ad09\") " Nov 23 20:11:58 crc kubenswrapper[4726]: I1123 20:11:58.905719 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-utilities" (OuterVolumeSpecName: "utilities") pod "0511e0ad-18eb-4add-9831-35762533ad09" (UID: "0511e0ad-18eb-4add-9831-35762533ad09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:11:58 crc kubenswrapper[4726]: I1123 20:11:58.922064 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0511e0ad-18eb-4add-9831-35762533ad09-kube-api-access-fcdlf" (OuterVolumeSpecName: "kube-api-access-fcdlf") pod "0511e0ad-18eb-4add-9831-35762533ad09" (UID: "0511e0ad-18eb-4add-9831-35762533ad09"). InnerVolumeSpecName "kube-api-access-fcdlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.006648 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0511e0ad-18eb-4add-9831-35762533ad09" (UID: "0511e0ad-18eb-4add-9831-35762533ad09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.006778 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.006791 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0511e0ad-18eb-4add-9831-35762533ad09-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.006801 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcdlf\" (UniqueName: \"kubernetes.io/projected/0511e0ad-18eb-4add-9831-35762533ad09-kube-api-access-fcdlf\") on node \"crc\" DevicePath \"\"" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.063542 4726 generic.go:334] "Generic (PLEG): container finished" podID="0511e0ad-18eb-4add-9831-35762533ad09" containerID="7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0" exitCode=0 Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.063753 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8jrp" event={"ID":"0511e0ad-18eb-4add-9831-35762533ad09","Type":"ContainerDied","Data":"7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0"} Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.063854 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8jrp" event={"ID":"0511e0ad-18eb-4add-9831-35762533ad09","Type":"ContainerDied","Data":"4b09e2a01d6b1a1698eb0dc024c801e68d7aaa5256295873a9c658787da4aec5"} Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.063941 4726 scope.go:117] "RemoveContainer" containerID="7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.064107 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8jrp" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.089084 4726 scope.go:117] "RemoveContainer" containerID="8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.091143 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8jrp"] Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.094361 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q8jrp"] Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.121429 4726 scope.go:117] "RemoveContainer" containerID="31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.135572 4726 scope.go:117] "RemoveContainer" containerID="7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0" Nov 23 20:11:59 crc kubenswrapper[4726]: E1123 20:11:59.136196 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0\": container with ID starting with 7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0 not found: ID does not exist" containerID="7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.136243 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0"} err="failed to get container status \"7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0\": rpc error: code = NotFound desc = could not find container \"7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0\": container with ID starting with 7d091f7cecfd5e835fe3f35ae5e087537f53d05afadc4a49885f37647eab63f0 not found: ID does not exist" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.136273 4726 scope.go:117] "RemoveContainer" containerID="8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f" Nov 23 20:11:59 crc kubenswrapper[4726]: E1123 20:11:59.136688 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f\": container with ID starting with 8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f not found: ID does not exist" containerID="8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.136732 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f"} err="failed to get container status \"8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f\": rpc error: code = NotFound desc = could not find container \"8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f\": container with ID starting with 8b867799ae03d420aea16b605d1659910ceea1e2ea0a35963fa97d0e057afc2f not found: ID does not exist" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.136763 4726 scope.go:117] "RemoveContainer" containerID="31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a" Nov 23 20:11:59 crc kubenswrapper[4726]: E1123 20:11:59.137215 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a\": container with ID starting with 31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a not found: ID does not exist" containerID="31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a" Nov 23 20:11:59 crc kubenswrapper[4726]: I1123 20:11:59.137240 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a"} err="failed to get container status \"31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a\": rpc error: code = NotFound desc = could not find container \"31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a\": container with ID starting with 31195d6189e7e2924e3f03cf4b6f8d30e74ebd231304873ac0666c8bff0d897a not found: ID does not exist" Nov 23 20:12:00 crc kubenswrapper[4726]: I1123 20:12:00.602408 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0511e0ad-18eb-4add-9831-35762533ad09" path="/var/lib/kubelet/pods/0511e0ad-18eb-4add-9831-35762533ad09/volumes" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.418980 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h4jxc"] Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.419786 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h4jxc" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" containerName="registry-server" containerID="cri-o://78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45" gracePeriod=30 Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.432537 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7csf"] Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.432782 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t7csf" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerName="registry-server" containerID="cri-o://28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c" gracePeriod=30 Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.444953 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mm77f"] Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.445147 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" containerName="marketplace-operator" containerID="cri-o://b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3" gracePeriod=30 Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.455526 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nlztb"] Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.455759 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nlztb" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="registry-server" containerID="cri-o://da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712" gracePeriod=30 Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.467328 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nc5vs"] Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.467535 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nc5vs" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerName="registry-server" containerID="cri-o://7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9" gracePeriod=30 Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.470359 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rvp2"] Nov 23 20:12:14 crc kubenswrapper[4726]: E1123 20:12:14.470517 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="extract-utilities" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.470533 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="extract-utilities" Nov 23 20:12:14 crc kubenswrapper[4726]: E1123 20:12:14.470548 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="registry-server" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.470554 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="registry-server" Nov 23 20:12:14 crc kubenswrapper[4726]: E1123 20:12:14.470567 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="extract-content" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.470573 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="extract-content" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.470650 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0511e0ad-18eb-4add-9831-35762533ad09" containerName="registry-server" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.470999 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.494811 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rvp2"] Nov 23 20:12:14 crc kubenswrapper[4726]: E1123 20:12:14.518504 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712 is running failed: container process not found" containerID="da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712" cmd=["grpc_health_probe","-addr=:50051"] Nov 23 20:12:14 crc kubenswrapper[4726]: E1123 20:12:14.518721 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712 is running failed: container process not found" containerID="da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712" cmd=["grpc_health_probe","-addr=:50051"] Nov 23 20:12:14 crc kubenswrapper[4726]: E1123 20:12:14.518894 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712 is running failed: container process not found" containerID="da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712" cmd=["grpc_health_probe","-addr=:50051"] Nov 23 20:12:14 crc kubenswrapper[4726]: E1123 20:12:14.518932 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-nlztb" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="registry-server" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.575442 4726 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mm77f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.575504 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.672952 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.672996 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.673034 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6jsc\" (UniqueName: \"kubernetes.io/projected/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-kube-api-access-n6jsc\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.773824 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.774168 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.774209 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6jsc\" (UniqueName: \"kubernetes.io/projected/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-kube-api-access-n6jsc\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.777775 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.789923 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.797578 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6jsc\" (UniqueName: \"kubernetes.io/projected/cfd53c53-6459-4a5d-b8e4-bf47ebb6da34-kube-api-access-n6jsc\") pod \"marketplace-operator-79b997595-7rvp2\" (UID: \"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.815000 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:12:14 crc kubenswrapper[4726]: E1123 20:12:14.830580 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15f81f72_746c_479f_b5b0_ba9ff588717a.slice/crio-28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c.scope\": RecentStats: unable to find data in memory cache]" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.863207 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.874842 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.888378 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976247 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-utilities\") pod \"6a8b32c1-af97-4361-ab85-75d154b38a76\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976302 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwqv2\" (UniqueName: \"kubernetes.io/projected/f8a59726-54f0-4469-b7fc-242607f33e83-kube-api-access-cwqv2\") pod \"f8a59726-54f0-4469-b7fc-242607f33e83\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976327 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-catalog-content\") pod \"f1364754-8bfa-45e8-8520-893acdb2837e\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976388 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stgr5\" (UniqueName: \"kubernetes.io/projected/f1364754-8bfa-45e8-8520-893acdb2837e-kube-api-access-stgr5\") pod \"f1364754-8bfa-45e8-8520-893acdb2837e\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976409 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw7md\" (UniqueName: \"kubernetes.io/projected/6a8b32c1-af97-4361-ab85-75d154b38a76-kube-api-access-qw7md\") pod \"6a8b32c1-af97-4361-ab85-75d154b38a76\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976443 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-trusted-ca\") pod \"f8a59726-54f0-4469-b7fc-242607f33e83\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976465 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-catalog-content\") pod \"6a8b32c1-af97-4361-ab85-75d154b38a76\" (UID: \"6a8b32c1-af97-4361-ab85-75d154b38a76\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976497 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-operator-metrics\") pod \"f8a59726-54f0-4469-b7fc-242607f33e83\" (UID: \"f8a59726-54f0-4469-b7fc-242607f33e83\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.976522 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-utilities\") pod \"f1364754-8bfa-45e8-8520-893acdb2837e\" (UID: \"f1364754-8bfa-45e8-8520-893acdb2837e\") " Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.977855 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-utilities" (OuterVolumeSpecName: "utilities") pod "6a8b32c1-af97-4361-ab85-75d154b38a76" (UID: "6a8b32c1-af97-4361-ab85-75d154b38a76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.978518 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-utilities" (OuterVolumeSpecName: "utilities") pod "f1364754-8bfa-45e8-8520-893acdb2837e" (UID: "f1364754-8bfa-45e8-8520-893acdb2837e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.980896 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f8a59726-54f0-4469-b7fc-242607f33e83" (UID: "f8a59726-54f0-4469-b7fc-242607f33e83"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.981050 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8b32c1-af97-4361-ab85-75d154b38a76-kube-api-access-qw7md" (OuterVolumeSpecName: "kube-api-access-qw7md") pod "6a8b32c1-af97-4361-ab85-75d154b38a76" (UID: "6a8b32c1-af97-4361-ab85-75d154b38a76"). InnerVolumeSpecName "kube-api-access-qw7md". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.983926 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f8a59726-54f0-4469-b7fc-242607f33e83" (UID: "f8a59726-54f0-4469-b7fc-242607f33e83"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.990531 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1364754-8bfa-45e8-8520-893acdb2837e-kube-api-access-stgr5" (OuterVolumeSpecName: "kube-api-access-stgr5") pod "f1364754-8bfa-45e8-8520-893acdb2837e" (UID: "f1364754-8bfa-45e8-8520-893acdb2837e"). InnerVolumeSpecName "kube-api-access-stgr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:12:14 crc kubenswrapper[4726]: I1123 20:12:14.995813 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8a59726-54f0-4469-b7fc-242607f33e83-kube-api-access-cwqv2" (OuterVolumeSpecName: "kube-api-access-cwqv2") pod "f8a59726-54f0-4469-b7fc-242607f33e83" (UID: "f8a59726-54f0-4469-b7fc-242607f33e83"). InnerVolumeSpecName "kube-api-access-cwqv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.006773 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.061179 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1364754-8bfa-45e8-8520-893acdb2837e" (UID: "f1364754-8bfa-45e8-8520-893acdb2837e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.077235 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlwm6\" (UniqueName: \"kubernetes.io/projected/15f81f72-746c-479f-b5b0-ba9ff588717a-kube-api-access-wlwm6\") pod \"15f81f72-746c-479f-b5b0-ba9ff588717a\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.077277 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-utilities\") pod \"15f81f72-746c-479f-b5b0-ba9ff588717a\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.077973 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-utilities" (OuterVolumeSpecName: "utilities") pod "15f81f72-746c-479f-b5b0-ba9ff588717a" (UID: "15f81f72-746c-479f-b5b0-ba9ff588717a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.078296 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-catalog-content\") pod \"15f81f72-746c-479f-b5b0-ba9ff588717a\" (UID: \"15f81f72-746c-479f-b5b0-ba9ff588717a\") " Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083509 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083560 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8a59726-54f0-4469-b7fc-242607f33e83-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083577 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083590 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083602 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwqv2\" (UniqueName: \"kubernetes.io/projected/f8a59726-54f0-4469-b7fc-242607f33e83-kube-api-access-cwqv2\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083619 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1364754-8bfa-45e8-8520-893acdb2837e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083628 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083637 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stgr5\" (UniqueName: \"kubernetes.io/projected/f1364754-8bfa-45e8-8520-893acdb2837e-kube-api-access-stgr5\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083647 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw7md\" (UniqueName: \"kubernetes.io/projected/6a8b32c1-af97-4361-ab85-75d154b38a76-kube-api-access-qw7md\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.083202 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15f81f72-746c-479f-b5b0-ba9ff588717a-kube-api-access-wlwm6" (OuterVolumeSpecName: "kube-api-access-wlwm6") pod "15f81f72-746c-479f-b5b0-ba9ff588717a" (UID: "15f81f72-746c-479f-b5b0-ba9ff588717a"). InnerVolumeSpecName "kube-api-access-wlwm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.094636 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.103339 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a8b32c1-af97-4361-ab85-75d154b38a76" (UID: "6a8b32c1-af97-4361-ab85-75d154b38a76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.142114 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15f81f72-746c-479f-b5b0-ba9ff588717a" (UID: "15f81f72-746c-479f-b5b0-ba9ff588717a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.154617 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1364754-8bfa-45e8-8520-893acdb2837e" containerID="78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45" exitCode=0 Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.154738 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4jxc" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.155791 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4jxc" event={"ID":"f1364754-8bfa-45e8-8520-893acdb2837e","Type":"ContainerDied","Data":"78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.156004 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4jxc" event={"ID":"f1364754-8bfa-45e8-8520-893acdb2837e","Type":"ContainerDied","Data":"dce19ccdb807c181cb4d618ce2ead2bc30290cb5137bf2b201a27a40394f4a7c"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.156043 4726 scope.go:117] "RemoveContainer" containerID="78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.161009 4726 generic.go:334] "Generic (PLEG): container finished" podID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerID="28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c" exitCode=0 Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.161174 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7csf" event={"ID":"15f81f72-746c-479f-b5b0-ba9ff588717a","Type":"ContainerDied","Data":"28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.161276 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7csf" event={"ID":"15f81f72-746c-479f-b5b0-ba9ff588717a","Type":"ContainerDied","Data":"c06383edbaff76b6202a069b11b6ef87f3c81096614f643967afd33a6c88b047"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.161226 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7csf" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.164223 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerID="7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9" exitCode=0 Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.164284 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nc5vs" event={"ID":"6a8b32c1-af97-4361-ab85-75d154b38a76","Type":"ContainerDied","Data":"7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.164311 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nc5vs" event={"ID":"6a8b32c1-af97-4361-ab85-75d154b38a76","Type":"ContainerDied","Data":"cae1f031d0f579370b5886a9f24d28d159c092e1e4c624e29af9d4053e7d2e0b"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.164396 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nc5vs" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.172546 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerID="da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712" exitCode=0 Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.172815 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nlztb" event={"ID":"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5","Type":"ContainerDied","Data":"da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.172849 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nlztb" event={"ID":"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5","Type":"ContainerDied","Data":"de25986a85e84e1791604d5788a5426d8ef374a3e9f803fd4faf0fc77b5b395d"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.172964 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nlztb" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.184142 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-catalog-content\") pod \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.184181 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-utilities\") pod \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.184287 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8kbr\" (UniqueName: \"kubernetes.io/projected/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-kube-api-access-r8kbr\") pod \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\" (UID: \"a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5\") " Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.184447 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlwm6\" (UniqueName: \"kubernetes.io/projected/15f81f72-746c-479f-b5b0-ba9ff588717a-kube-api-access-wlwm6\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.184458 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f81f72-746c-479f-b5b0-ba9ff588717a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.184467 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b32c1-af97-4361-ab85-75d154b38a76-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.188229 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.188172 4726 generic.go:334] "Generic (PLEG): container finished" podID="f8a59726-54f0-4469-b7fc-242607f33e83" containerID="b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3" exitCode=0 Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.188700 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" event={"ID":"f8a59726-54f0-4469-b7fc-242607f33e83","Type":"ContainerDied","Data":"b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.188761 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mm77f" event={"ID":"f8a59726-54f0-4469-b7fc-242607f33e83","Type":"ContainerDied","Data":"49d0b24032c0976f1a239fd9d4023798ecf0dd4cd428cd11e8b5562e1cb7aaf8"} Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.189004 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-utilities" (OuterVolumeSpecName: "utilities") pod "a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" (UID: "a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.198129 4726 scope.go:117] "RemoveContainer" containerID="29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.203944 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nc5vs"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.206349 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-kube-api-access-r8kbr" (OuterVolumeSpecName: "kube-api-access-r8kbr") pod "a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" (UID: "a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5"). InnerVolumeSpecName "kube-api-access-r8kbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.208366 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nc5vs"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.221878 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h4jxc"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.237836 4726 scope.go:117] "RemoveContainer" containerID="67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.240500 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" (UID: "a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.241213 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h4jxc"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.251375 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mm77f"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.259489 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mm77f"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.266225 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7csf"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.269460 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t7csf"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.276252 4726 scope.go:117] "RemoveContainer" containerID="78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.276728 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45\": container with ID starting with 78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45 not found: ID does not exist" containerID="78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.276771 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45"} err="failed to get container status \"78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45\": rpc error: code = NotFound desc = could not find container \"78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45\": container with ID starting with 78494758fc8cb16753717fd6050236e741ff8fc00445dd098d05e7345fb97f45 not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.276943 4726 scope.go:117] "RemoveContainer" containerID="29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.277460 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0\": container with ID starting with 29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0 not found: ID does not exist" containerID="29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.277496 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0"} err="failed to get container status \"29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0\": rpc error: code = NotFound desc = could not find container \"29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0\": container with ID starting with 29b4a741b5e817e6fcacf9fafd43f70f26858d953452208acfb7b6b6069685d0 not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.277518 4726 scope.go:117] "RemoveContainer" containerID="67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.277862 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375\": container with ID starting with 67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375 not found: ID does not exist" containerID="67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.277895 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375"} err="failed to get container status \"67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375\": rpc error: code = NotFound desc = could not find container \"67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375\": container with ID starting with 67e19b199e5950365af4886001d0fc1d3ba0e90a1e9db19456013ade72e09375 not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.277908 4726 scope.go:117] "RemoveContainer" containerID="28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.285991 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8kbr\" (UniqueName: \"kubernetes.io/projected/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-kube-api-access-r8kbr\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.286024 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.286036 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.289597 4726 scope.go:117] "RemoveContainer" containerID="692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.301452 4726 scope.go:117] "RemoveContainer" containerID="285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.317169 4726 scope.go:117] "RemoveContainer" containerID="28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.317565 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c\": container with ID starting with 28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c not found: ID does not exist" containerID="28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.317592 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c"} err="failed to get container status \"28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c\": rpc error: code = NotFound desc = could not find container \"28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c\": container with ID starting with 28d2b159e7a31a931d72f8c3a7b5bef6ac4f294580cbfd71a0c05562555a421c not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.317614 4726 scope.go:117] "RemoveContainer" containerID="692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.318051 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212\": container with ID starting with 692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212 not found: ID does not exist" containerID="692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.318071 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212"} err="failed to get container status \"692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212\": rpc error: code = NotFound desc = could not find container \"692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212\": container with ID starting with 692b7e32ab77c83ba10142d1952fe9a59e194406ed67c771194abc3232923212 not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.318084 4726 scope.go:117] "RemoveContainer" containerID="285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.318357 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c\": container with ID starting with 285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c not found: ID does not exist" containerID="285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.318379 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c"} err="failed to get container status \"285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c\": rpc error: code = NotFound desc = could not find container \"285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c\": container with ID starting with 285a7eede2bbe23e576f6daa7d811faf93c09db8ad438478fe8bd3f73f9e543c not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.318391 4726 scope.go:117] "RemoveContainer" containerID="7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.330703 4726 scope.go:117] "RemoveContainer" containerID="d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.347272 4726 scope.go:117] "RemoveContainer" containerID="f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.360052 4726 scope.go:117] "RemoveContainer" containerID="7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.360438 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9\": container with ID starting with 7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9 not found: ID does not exist" containerID="7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.360467 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9"} err="failed to get container status \"7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9\": rpc error: code = NotFound desc = could not find container \"7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9\": container with ID starting with 7f6244c8023491a6299847d3ce3c8d4c05b594aa0fa3981f66d1ca2147131fb9 not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.360487 4726 scope.go:117] "RemoveContainer" containerID="d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.360941 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e\": container with ID starting with d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e not found: ID does not exist" containerID="d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.360978 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e"} err="failed to get container status \"d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e\": rpc error: code = NotFound desc = could not find container \"d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e\": container with ID starting with d998c928e99c70c7365a7e8d6ad7f6298d28c0504fc475ef3a018e811c21d93e not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.361026 4726 scope.go:117] "RemoveContainer" containerID="f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.361406 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df\": container with ID starting with f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df not found: ID does not exist" containerID="f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.361431 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df"} err="failed to get container status \"f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df\": rpc error: code = NotFound desc = could not find container \"f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df\": container with ID starting with f3e0bfc104aa55d1ca45204cf48abb9d491c354c8d01110a134996b3e552d4df not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.361445 4726 scope.go:117] "RemoveContainer" containerID="da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.372680 4726 scope.go:117] "RemoveContainer" containerID="44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.384116 4726 scope.go:117] "RemoveContainer" containerID="9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.396221 4726 scope.go:117] "RemoveContainer" containerID="da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.396683 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712\": container with ID starting with da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712 not found: ID does not exist" containerID="da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.396754 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712"} err="failed to get container status \"da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712\": rpc error: code = NotFound desc = could not find container \"da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712\": container with ID starting with da8cd31bf7e8bbe085b619ac9878b4818c3f76c72d6f9fab5175c61c825c2712 not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.396778 4726 scope.go:117] "RemoveContainer" containerID="44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.397168 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949\": container with ID starting with 44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949 not found: ID does not exist" containerID="44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.397193 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949"} err="failed to get container status \"44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949\": rpc error: code = NotFound desc = could not find container \"44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949\": container with ID starting with 44d829f233e9d7cc26b67f4647c80625e3620742758b34a538f054256f390949 not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.397244 4726 scope.go:117] "RemoveContainer" containerID="9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.397756 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae\": container with ID starting with 9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae not found: ID does not exist" containerID="9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.397823 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae"} err="failed to get container status \"9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae\": rpc error: code = NotFound desc = could not find container \"9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae\": container with ID starting with 9f5ebee8617dd0c07654328741a8fd32e1a96eb575b8ae09ef7c4cebf7ff03ae not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.397838 4726 scope.go:117] "RemoveContainer" containerID="b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.409834 4726 scope.go:117] "RemoveContainer" containerID="b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3" Nov 23 20:12:15 crc kubenswrapper[4726]: E1123 20:12:15.410319 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3\": container with ID starting with b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3 not found: ID does not exist" containerID="b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.410347 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3"} err="failed to get container status \"b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3\": rpc error: code = NotFound desc = could not find container \"b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3\": container with ID starting with b3f31ca66a4f781ace79753f52a4dd784d6e4133d7d95aea5a5f32197b6dede3 not found: ID does not exist" Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.486604 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rvp2"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.507053 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nlztb"] Nov 23 20:12:15 crc kubenswrapper[4726]: I1123 20:12:15.508838 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nlztb"] Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.198106 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" event={"ID":"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34","Type":"ContainerStarted","Data":"0b8ef29bccbeefbc2d5a0f17ba5fec1cb3f02f3617514ec1b9879db279e01462"} Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.198460 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" event={"ID":"cfd53c53-6459-4a5d-b8e4-bf47ebb6da34","Type":"ContainerStarted","Data":"8bddf08c87f9af5b50a41a2b26e36440d7f96be9659fd337009d27bad8f55e49"} Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.199391 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.225191 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.243914 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7rvp2" podStartSLOduration=2.24389635 podStartE2EDuration="2.24389635s" podCreationTimestamp="2025-11-23 20:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:12:16.23920842 +0000 UTC m=+244.388249396" watchObservedRunningTime="2025-11-23 20:12:16.24389635 +0000 UTC m=+244.392937306" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.595493 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" path="/var/lib/kubelet/pods/15f81f72-746c-479f-b5b0-ba9ff588717a/volumes" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.596184 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" path="/var/lib/kubelet/pods/6a8b32c1-af97-4361-ab85-75d154b38a76/volumes" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.596745 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" path="/var/lib/kubelet/pods/a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5/volumes" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.597385 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" path="/var/lib/kubelet/pods/f1364754-8bfa-45e8-8520-893acdb2837e/volumes" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.598706 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" path="/var/lib/kubelet/pods/f8a59726-54f0-4469-b7fc-242607f33e83/volumes" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919220 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b28ts"] Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919442 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="extract-utilities" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919457 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="extract-utilities" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919467 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" containerName="marketplace-operator" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919474 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" containerName="marketplace-operator" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919485 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919493 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919502 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerName="extract-content" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919510 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerName="extract-content" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919537 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" containerName="extract-content" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919545 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" containerName="extract-content" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919556 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerName="extract-utilities" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919562 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerName="extract-utilities" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919570 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="extract-content" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919576 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="extract-content" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919585 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919590 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919598 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerName="extract-utilities" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919604 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerName="extract-utilities" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919611 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" containerName="extract-utilities" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919616 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" containerName="extract-utilities" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919625 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerName="extract-content" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919630 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerName="extract-content" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919637 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919643 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: E1123 20:12:16.919649 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919655 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919730 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1364754-8bfa-45e8-8520-893acdb2837e" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919741 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8a59726-54f0-4469-b7fc-242607f33e83" containerName="marketplace-operator" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919751 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8b32c1-af97-4361-ab85-75d154b38a76" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919759 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c99d42-1286-4b5b-8fa2-8673a0bcc2a5" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.919769 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f81f72-746c-479f-b5b0-ba9ff588717a" containerName="registry-server" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.920414 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.927612 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 23 20:12:16 crc kubenswrapper[4726]: I1123 20:12:16.930364 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b28ts"] Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.108643 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75ecd57d-a57d-44fc-811a-1d30d925e93e-utilities\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.108686 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bc99\" (UniqueName: \"kubernetes.io/projected/75ecd57d-a57d-44fc-811a-1d30d925e93e-kube-api-access-8bc99\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.108714 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75ecd57d-a57d-44fc-811a-1d30d925e93e-catalog-content\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.209654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75ecd57d-a57d-44fc-811a-1d30d925e93e-utilities\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.209682 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bc99\" (UniqueName: \"kubernetes.io/projected/75ecd57d-a57d-44fc-811a-1d30d925e93e-kube-api-access-8bc99\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.209808 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75ecd57d-a57d-44fc-811a-1d30d925e93e-catalog-content\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.210342 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75ecd57d-a57d-44fc-811a-1d30d925e93e-utilities\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.210486 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75ecd57d-a57d-44fc-811a-1d30d925e93e-catalog-content\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.226986 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bc99\" (UniqueName: \"kubernetes.io/projected/75ecd57d-a57d-44fc-811a-1d30d925e93e-kube-api-access-8bc99\") pod \"redhat-marketplace-b28ts\" (UID: \"75ecd57d-a57d-44fc-811a-1d30d925e93e\") " pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.238486 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.526170 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t8776"] Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.532943 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.535798 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.541479 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t8776"] Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.615069 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5073784-d975-4fce-805f-7216d9356df8-utilities\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.616171 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5073784-d975-4fce-805f-7216d9356df8-catalog-content\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.616289 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2wlb\" (UniqueName: \"kubernetes.io/projected/d5073784-d975-4fce-805f-7216d9356df8-kube-api-access-k2wlb\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.660159 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b28ts"] Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.718233 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5073784-d975-4fce-805f-7216d9356df8-utilities\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.718601 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5073784-d975-4fce-805f-7216d9356df8-catalog-content\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.718684 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2wlb\" (UniqueName: \"kubernetes.io/projected/d5073784-d975-4fce-805f-7216d9356df8-kube-api-access-k2wlb\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.719168 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5073784-d975-4fce-805f-7216d9356df8-utilities\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.719650 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5073784-d975-4fce-805f-7216d9356df8-catalog-content\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.735333 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2wlb\" (UniqueName: \"kubernetes.io/projected/d5073784-d975-4fce-805f-7216d9356df8-kube-api-access-k2wlb\") pod \"redhat-operators-t8776\" (UID: \"d5073784-d975-4fce-805f-7216d9356df8\") " pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:17 crc kubenswrapper[4726]: I1123 20:12:17.845307 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:18 crc kubenswrapper[4726]: I1123 20:12:18.214073 4726 generic.go:334] "Generic (PLEG): container finished" podID="75ecd57d-a57d-44fc-811a-1d30d925e93e" containerID="4a4a17ad28c9879d3d5c51e3ba5df81282e87939629fd3c4b3da7c1f8cab02fa" exitCode=0 Nov 23 20:12:18 crc kubenswrapper[4726]: I1123 20:12:18.214195 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b28ts" event={"ID":"75ecd57d-a57d-44fc-811a-1d30d925e93e","Type":"ContainerDied","Data":"4a4a17ad28c9879d3d5c51e3ba5df81282e87939629fd3c4b3da7c1f8cab02fa"} Nov 23 20:12:18 crc kubenswrapper[4726]: I1123 20:12:18.214351 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b28ts" event={"ID":"75ecd57d-a57d-44fc-811a-1d30d925e93e","Type":"ContainerStarted","Data":"3ca951062f5e51b801da94934b169dd4c1afd957ec1c02588d72030225b7d523"} Nov 23 20:12:18 crc kubenswrapper[4726]: I1123 20:12:18.272848 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t8776"] Nov 23 20:12:18 crc kubenswrapper[4726]: W1123 20:12:18.283372 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5073784_d975_4fce_805f_7216d9356df8.slice/crio-eeb1ffbaa473a6ebc01c74341c5c76d1260817f2c421744212a55a68d162755c WatchSource:0}: Error finding container eeb1ffbaa473a6ebc01c74341c5c76d1260817f2c421744212a55a68d162755c: Status 404 returned error can't find the container with id eeb1ffbaa473a6ebc01c74341c5c76d1260817f2c421744212a55a68d162755c Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.220832 4726 generic.go:334] "Generic (PLEG): container finished" podID="d5073784-d975-4fce-805f-7216d9356df8" containerID="4378d37304c07f10e516619740884a1b9120e4a6f9b7b396d283439a0acb207a" exitCode=0 Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.220932 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8776" event={"ID":"d5073784-d975-4fce-805f-7216d9356df8","Type":"ContainerDied","Data":"4378d37304c07f10e516619740884a1b9120e4a6f9b7b396d283439a0acb207a"} Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.221018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8776" event={"ID":"d5073784-d975-4fce-805f-7216d9356df8","Type":"ContainerStarted","Data":"eeb1ffbaa473a6ebc01c74341c5c76d1260817f2c421744212a55a68d162755c"} Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.223807 4726 generic.go:334] "Generic (PLEG): container finished" podID="75ecd57d-a57d-44fc-811a-1d30d925e93e" containerID="fa04453b0a4c0f9959619ae402e989c80e5198ebe1a84d4b1b37e6de86369e64" exitCode=0 Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.223851 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b28ts" event={"ID":"75ecd57d-a57d-44fc-811a-1d30d925e93e","Type":"ContainerDied","Data":"fa04453b0a4c0f9959619ae402e989c80e5198ebe1a84d4b1b37e6de86369e64"} Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.327264 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8qfs7"] Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.328820 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.331674 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.335976 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8qfs7"] Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.350191 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-utilities\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.350254 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-catalog-content\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.350360 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpm9w\" (UniqueName: \"kubernetes.io/projected/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-kube-api-access-vpm9w\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.451115 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpm9w\" (UniqueName: \"kubernetes.io/projected/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-kube-api-access-vpm9w\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.451155 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-utilities\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.451175 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-catalog-content\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.451583 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-catalog-content\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.451731 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-utilities\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.474078 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpm9w\" (UniqueName: \"kubernetes.io/projected/5bb09cfa-b62c-4b90-9cbc-2309be53c15a-kube-api-access-vpm9w\") pod \"certified-operators-8qfs7\" (UID: \"5bb09cfa-b62c-4b90-9cbc-2309be53c15a\") " pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.655849 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.922147 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r7nwg"] Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.923381 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.928294 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.935886 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r7nwg"] Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.957583 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7459f94c-2ed1-470a-a308-1377af995ab7-utilities\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.957633 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7459f94c-2ed1-470a-a308-1377af995ab7-catalog-content\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:19 crc kubenswrapper[4726]: I1123 20:12:19.957845 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtgwg\" (UniqueName: \"kubernetes.io/projected/7459f94c-2ed1-470a-a308-1377af995ab7-kube-api-access-vtgwg\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.058633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7459f94c-2ed1-470a-a308-1377af995ab7-utilities\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.058679 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7459f94c-2ed1-470a-a308-1377af995ab7-catalog-content\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.058744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtgwg\" (UniqueName: \"kubernetes.io/projected/7459f94c-2ed1-470a-a308-1377af995ab7-kube-api-access-vtgwg\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.059168 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7459f94c-2ed1-470a-a308-1377af995ab7-catalog-content\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.059180 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7459f94c-2ed1-470a-a308-1377af995ab7-utilities\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.076438 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtgwg\" (UniqueName: \"kubernetes.io/projected/7459f94c-2ed1-470a-a308-1377af995ab7-kube-api-access-vtgwg\") pod \"community-operators-r7nwg\" (UID: \"7459f94c-2ed1-470a-a308-1377af995ab7\") " pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.111807 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8qfs7"] Nov 23 20:12:20 crc kubenswrapper[4726]: W1123 20:12:20.118608 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bb09cfa_b62c_4b90_9cbc_2309be53c15a.slice/crio-da63fd1137f5d600b60242ee4aff29c19367ac0f15c4d516d6811af0969d2f30 WatchSource:0}: Error finding container da63fd1137f5d600b60242ee4aff29c19367ac0f15c4d516d6811af0969d2f30: Status 404 returned error can't find the container with id da63fd1137f5d600b60242ee4aff29c19367ac0f15c4d516d6811af0969d2f30 Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.231367 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b28ts" event={"ID":"75ecd57d-a57d-44fc-811a-1d30d925e93e","Type":"ContainerStarted","Data":"97b52b21311bc7ebf56686573f79f530e21dc9d0ba3a8cb9c95ca4fa85384dd6"} Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.234508 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfs7" event={"ID":"5bb09cfa-b62c-4b90-9cbc-2309be53c15a","Type":"ContainerStarted","Data":"91ebc13ede32f37bf293d576aa492139e28a0619d412dc5ace6d7f19f31d959d"} Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.234570 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfs7" event={"ID":"5bb09cfa-b62c-4b90-9cbc-2309be53c15a","Type":"ContainerStarted","Data":"da63fd1137f5d600b60242ee4aff29c19367ac0f15c4d516d6811af0969d2f30"} Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.239372 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8776" event={"ID":"d5073784-d975-4fce-805f-7216d9356df8","Type":"ContainerStarted","Data":"dbc8458e722d8c2c9048661f3e1e4a5cbc65d509a338ba85ebb4098313c573c6"} Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.243346 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.247020 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b28ts" podStartSLOduration=2.751465092 podStartE2EDuration="4.247011273s" podCreationTimestamp="2025-11-23 20:12:16 +0000 UTC" firstStartedPulling="2025-11-23 20:12:18.216777961 +0000 UTC m=+246.365818927" lastFinishedPulling="2025-11-23 20:12:19.712324162 +0000 UTC m=+247.861365108" observedRunningTime="2025-11-23 20:12:20.245078785 +0000 UTC m=+248.394119741" watchObservedRunningTime="2025-11-23 20:12:20.247011273 +0000 UTC m=+248.396052229" Nov 23 20:12:20 crc kubenswrapper[4726]: I1123 20:12:20.661498 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r7nwg"] Nov 23 20:12:20 crc kubenswrapper[4726]: W1123 20:12:20.674609 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7459f94c_2ed1_470a_a308_1377af995ab7.slice/crio-1a2414973fc6d8b402fd66a2cd055f6f829542c544cf7c6a81aab9d1ad968480 WatchSource:0}: Error finding container 1a2414973fc6d8b402fd66a2cd055f6f829542c544cf7c6a81aab9d1ad968480: Status 404 returned error can't find the container with id 1a2414973fc6d8b402fd66a2cd055f6f829542c544cf7c6a81aab9d1ad968480 Nov 23 20:12:21 crc kubenswrapper[4726]: I1123 20:12:21.245059 4726 generic.go:334] "Generic (PLEG): container finished" podID="5bb09cfa-b62c-4b90-9cbc-2309be53c15a" containerID="91ebc13ede32f37bf293d576aa492139e28a0619d412dc5ace6d7f19f31d959d" exitCode=0 Nov 23 20:12:21 crc kubenswrapper[4726]: I1123 20:12:21.245128 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfs7" event={"ID":"5bb09cfa-b62c-4b90-9cbc-2309be53c15a","Type":"ContainerDied","Data":"91ebc13ede32f37bf293d576aa492139e28a0619d412dc5ace6d7f19f31d959d"} Nov 23 20:12:21 crc kubenswrapper[4726]: I1123 20:12:21.246831 4726 generic.go:334] "Generic (PLEG): container finished" podID="7459f94c-2ed1-470a-a308-1377af995ab7" containerID="7bf66e95c480a016ce05c5ef7d0e3dcad55a615def546eeaa7b0097a7f1ccf5c" exitCode=0 Nov 23 20:12:21 crc kubenswrapper[4726]: I1123 20:12:21.246900 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nwg" event={"ID":"7459f94c-2ed1-470a-a308-1377af995ab7","Type":"ContainerDied","Data":"7bf66e95c480a016ce05c5ef7d0e3dcad55a615def546eeaa7b0097a7f1ccf5c"} Nov 23 20:12:21 crc kubenswrapper[4726]: I1123 20:12:21.247079 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nwg" event={"ID":"7459f94c-2ed1-470a-a308-1377af995ab7","Type":"ContainerStarted","Data":"1a2414973fc6d8b402fd66a2cd055f6f829542c544cf7c6a81aab9d1ad968480"} Nov 23 20:12:21 crc kubenswrapper[4726]: I1123 20:12:21.250630 4726 generic.go:334] "Generic (PLEG): container finished" podID="d5073784-d975-4fce-805f-7216d9356df8" containerID="dbc8458e722d8c2c9048661f3e1e4a5cbc65d509a338ba85ebb4098313c573c6" exitCode=0 Nov 23 20:12:21 crc kubenswrapper[4726]: I1123 20:12:21.251236 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8776" event={"ID":"d5073784-d975-4fce-805f-7216d9356df8","Type":"ContainerDied","Data":"dbc8458e722d8c2c9048661f3e1e4a5cbc65d509a338ba85ebb4098313c573c6"} Nov 23 20:12:22 crc kubenswrapper[4726]: I1123 20:12:22.256568 4726 generic.go:334] "Generic (PLEG): container finished" podID="7459f94c-2ed1-470a-a308-1377af995ab7" containerID="b35a1546618902bf6ab89ed1a89ff574861ecf4545c089f958ffaab4490562c5" exitCode=0 Nov 23 20:12:22 crc kubenswrapper[4726]: I1123 20:12:22.256634 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nwg" event={"ID":"7459f94c-2ed1-470a-a308-1377af995ab7","Type":"ContainerDied","Data":"b35a1546618902bf6ab89ed1a89ff574861ecf4545c089f958ffaab4490562c5"} Nov 23 20:12:22 crc kubenswrapper[4726]: I1123 20:12:22.259606 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8776" event={"ID":"d5073784-d975-4fce-805f-7216d9356df8","Type":"ContainerStarted","Data":"1c01ad9077b74e3f86fb8fde8b400a1f5320bbb793d8558daf0162fa16f15097"} Nov 23 20:12:22 crc kubenswrapper[4726]: I1123 20:12:22.262554 4726 generic.go:334] "Generic (PLEG): container finished" podID="5bb09cfa-b62c-4b90-9cbc-2309be53c15a" containerID="4fba554a56bc958b17a37edf3afc645bb1295735248c15208ca5ef7825c850af" exitCode=0 Nov 23 20:12:22 crc kubenswrapper[4726]: I1123 20:12:22.262580 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfs7" event={"ID":"5bb09cfa-b62c-4b90-9cbc-2309be53c15a","Type":"ContainerDied","Data":"4fba554a56bc958b17a37edf3afc645bb1295735248c15208ca5ef7825c850af"} Nov 23 20:12:22 crc kubenswrapper[4726]: I1123 20:12:22.307834 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t8776" podStartSLOduration=2.880804161 podStartE2EDuration="5.307815324s" podCreationTimestamp="2025-11-23 20:12:17 +0000 UTC" firstStartedPulling="2025-11-23 20:12:19.229778644 +0000 UTC m=+247.378819600" lastFinishedPulling="2025-11-23 20:12:21.656789807 +0000 UTC m=+249.805830763" observedRunningTime="2025-11-23 20:12:22.29390232 +0000 UTC m=+250.442943276" watchObservedRunningTime="2025-11-23 20:12:22.307815324 +0000 UTC m=+250.456856280" Nov 23 20:12:24 crc kubenswrapper[4726]: I1123 20:12:24.275966 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nwg" event={"ID":"7459f94c-2ed1-470a-a308-1377af995ab7","Type":"ContainerStarted","Data":"e0dee9fe9bf84843ce3f2f90a8612b5abc3ce5a7daa1ee49578cf25b7bbee988"} Nov 23 20:12:24 crc kubenswrapper[4726]: I1123 20:12:24.278249 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qfs7" event={"ID":"5bb09cfa-b62c-4b90-9cbc-2309be53c15a","Type":"ContainerStarted","Data":"a905cde84bd5ae7afe2ab0f2ca64e73549a621fb3559734470ee0e452a915e3a"} Nov 23 20:12:24 crc kubenswrapper[4726]: I1123 20:12:24.292494 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r7nwg" podStartSLOduration=3.851257577 podStartE2EDuration="5.292479498s" podCreationTimestamp="2025-11-23 20:12:19 +0000 UTC" firstStartedPulling="2025-11-23 20:12:21.248988217 +0000 UTC m=+249.398029173" lastFinishedPulling="2025-11-23 20:12:22.690210138 +0000 UTC m=+250.839251094" observedRunningTime="2025-11-23 20:12:24.291028975 +0000 UTC m=+252.440069931" watchObservedRunningTime="2025-11-23 20:12:24.292479498 +0000 UTC m=+252.441520454" Nov 23 20:12:27 crc kubenswrapper[4726]: I1123 20:12:27.240394 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:27 crc kubenswrapper[4726]: I1123 20:12:27.240707 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:27 crc kubenswrapper[4726]: I1123 20:12:27.277568 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:27 crc kubenswrapper[4726]: I1123 20:12:27.377529 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8qfs7" podStartSLOduration=5.85609105 podStartE2EDuration="8.377388762s" podCreationTimestamp="2025-11-23 20:12:19 +0000 UTC" firstStartedPulling="2025-11-23 20:12:20.235547991 +0000 UTC m=+248.384588947" lastFinishedPulling="2025-11-23 20:12:22.756845703 +0000 UTC m=+250.905886659" observedRunningTime="2025-11-23 20:12:24.313068841 +0000 UTC m=+252.462109797" watchObservedRunningTime="2025-11-23 20:12:27.377388762 +0000 UTC m=+255.526429718" Nov 23 20:12:27 crc kubenswrapper[4726]: I1123 20:12:27.410685 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b28ts" Nov 23 20:12:27 crc kubenswrapper[4726]: I1123 20:12:27.846575 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:27 crc kubenswrapper[4726]: I1123 20:12:27.846609 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:27 crc kubenswrapper[4726]: I1123 20:12:27.894075 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:28 crc kubenswrapper[4726]: I1123 20:12:28.361555 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t8776" Nov 23 20:12:29 crc kubenswrapper[4726]: I1123 20:12:29.656511 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:29 crc kubenswrapper[4726]: I1123 20:12:29.656940 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:29 crc kubenswrapper[4726]: I1123 20:12:29.710075 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:12:30 crc kubenswrapper[4726]: I1123 20:12:30.244284 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:30 crc kubenswrapper[4726]: I1123 20:12:30.244321 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:30 crc kubenswrapper[4726]: I1123 20:12:30.286618 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:30 crc kubenswrapper[4726]: I1123 20:12:30.355462 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r7nwg" Nov 23 20:12:30 crc kubenswrapper[4726]: I1123 20:12:30.393583 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8qfs7" Nov 23 20:13:39 crc kubenswrapper[4726]: I1123 20:13:39.043527 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:13:39 crc kubenswrapper[4726]: I1123 20:13:39.044408 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:14:09 crc kubenswrapper[4726]: I1123 20:14:09.043470 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:14:09 crc kubenswrapper[4726]: I1123 20:14:09.044356 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:14:39 crc kubenswrapper[4726]: I1123 20:14:39.043310 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:14:39 crc kubenswrapper[4726]: I1123 20:14:39.045097 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:14:39 crc kubenswrapper[4726]: I1123 20:14:39.045218 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:14:39 crc kubenswrapper[4726]: I1123 20:14:39.046173 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d69e63c1f53e98006b9aa5285330cc02cef0dd0c90e397e35c78c1a1ac967d6f"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:14:39 crc kubenswrapper[4726]: I1123 20:14:39.046271 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://d69e63c1f53e98006b9aa5285330cc02cef0dd0c90e397e35c78c1a1ac967d6f" gracePeriod=600 Nov 23 20:14:40 crc kubenswrapper[4726]: I1123 20:14:40.162983 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="d69e63c1f53e98006b9aa5285330cc02cef0dd0c90e397e35c78c1a1ac967d6f" exitCode=0 Nov 23 20:14:40 crc kubenswrapper[4726]: I1123 20:14:40.163052 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"d69e63c1f53e98006b9aa5285330cc02cef0dd0c90e397e35c78c1a1ac967d6f"} Nov 23 20:14:40 crc kubenswrapper[4726]: I1123 20:14:40.163922 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"5b42f4b47c86a8ff3e157d6e822c8e9bbf78256d76d3c430faf21561b98b7270"} Nov 23 20:14:40 crc kubenswrapper[4726]: I1123 20:14:40.163970 4726 scope.go:117] "RemoveContainer" containerID="396db3d62c2f860ead7633187c144a37b2459df6b644d4a2d0ff2ee5feb1ee60" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.149123 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj"] Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.152972 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.157986 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.160225 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.162479 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj"] Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.287564 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-secret-volume\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.287621 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnr6g\" (UniqueName: \"kubernetes.io/projected/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-kube-api-access-fnr6g\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.287665 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-config-volume\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.388419 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-secret-volume\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.388472 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnr6g\" (UniqueName: \"kubernetes.io/projected/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-kube-api-access-fnr6g\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.388516 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-config-volume\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.389640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-config-volume\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.395896 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-secret-volume\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.409140 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnr6g\" (UniqueName: \"kubernetes.io/projected/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-kube-api-access-fnr6g\") pod \"collect-profiles-29398815-bh5mj\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.470119 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:00 crc kubenswrapper[4726]: I1123 20:15:00.878484 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj"] Nov 23 20:15:00 crc kubenswrapper[4726]: W1123 20:15:00.892350 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6cef3ad9_bcdd_417e_85d1_972a8ed016a5.slice/crio-f5b1bc83d0e49cdb53620645dfd282e4d2ac8374f554609d4200e07600adad27 WatchSource:0}: Error finding container f5b1bc83d0e49cdb53620645dfd282e4d2ac8374f554609d4200e07600adad27: Status 404 returned error can't find the container with id f5b1bc83d0e49cdb53620645dfd282e4d2ac8374f554609d4200e07600adad27 Nov 23 20:15:01 crc kubenswrapper[4726]: I1123 20:15:01.303737 4726 generic.go:334] "Generic (PLEG): container finished" podID="6cef3ad9-bcdd-417e-85d1-972a8ed016a5" containerID="04f28e575b48de59825ef33c0d0c6f9a275917c035a138d90859f704bdd06126" exitCode=0 Nov 23 20:15:01 crc kubenswrapper[4726]: I1123 20:15:01.303784 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" event={"ID":"6cef3ad9-bcdd-417e-85d1-972a8ed016a5","Type":"ContainerDied","Data":"04f28e575b48de59825ef33c0d0c6f9a275917c035a138d90859f704bdd06126"} Nov 23 20:15:01 crc kubenswrapper[4726]: I1123 20:15:01.303810 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" event={"ID":"6cef3ad9-bcdd-417e-85d1-972a8ed016a5","Type":"ContainerStarted","Data":"f5b1bc83d0e49cdb53620645dfd282e4d2ac8374f554609d4200e07600adad27"} Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.555445 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.619185 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnr6g\" (UniqueName: \"kubernetes.io/projected/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-kube-api-access-fnr6g\") pod \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.619252 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-config-volume\") pod \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.619326 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-secret-volume\") pod \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\" (UID: \"6cef3ad9-bcdd-417e-85d1-972a8ed016a5\") " Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.619937 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-config-volume" (OuterVolumeSpecName: "config-volume") pod "6cef3ad9-bcdd-417e-85d1-972a8ed016a5" (UID: "6cef3ad9-bcdd-417e-85d1-972a8ed016a5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.620214 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.624598 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6cef3ad9-bcdd-417e-85d1-972a8ed016a5" (UID: "6cef3ad9-bcdd-417e-85d1-972a8ed016a5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.624598 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-kube-api-access-fnr6g" (OuterVolumeSpecName: "kube-api-access-fnr6g") pod "6cef3ad9-bcdd-417e-85d1-972a8ed016a5" (UID: "6cef3ad9-bcdd-417e-85d1-972a8ed016a5"). InnerVolumeSpecName "kube-api-access-fnr6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.721997 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:15:02 crc kubenswrapper[4726]: I1123 20:15:02.722028 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnr6g\" (UniqueName: \"kubernetes.io/projected/6cef3ad9-bcdd-417e-85d1-972a8ed016a5-kube-api-access-fnr6g\") on node \"crc\" DevicePath \"\"" Nov 23 20:15:03 crc kubenswrapper[4726]: I1123 20:15:03.317560 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" event={"ID":"6cef3ad9-bcdd-417e-85d1-972a8ed016a5","Type":"ContainerDied","Data":"f5b1bc83d0e49cdb53620645dfd282e4d2ac8374f554609d4200e07600adad27"} Nov 23 20:15:03 crc kubenswrapper[4726]: I1123 20:15:03.317810 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj" Nov 23 20:15:03 crc kubenswrapper[4726]: I1123 20:15:03.317825 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5b1bc83d0e49cdb53620645dfd282e4d2ac8374f554609d4200e07600adad27" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.080565 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z49ph"] Nov 23 20:16:38 crc kubenswrapper[4726]: E1123 20:16:38.081188 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cef3ad9-bcdd-417e-85d1-972a8ed016a5" containerName="collect-profiles" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.081199 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cef3ad9-bcdd-417e-85d1-972a8ed016a5" containerName="collect-profiles" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.081295 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cef3ad9-bcdd-417e-85d1-972a8ed016a5" containerName="collect-profiles" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.081649 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.124046 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z49ph"] Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.167394 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/321fd323-3561-4655-b218-3972b1d2d881-registry-certificates\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.167477 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-registry-tls\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.167583 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/321fd323-3561-4655-b218-3972b1d2d881-trusted-ca\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.167665 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/321fd323-3561-4655-b218-3972b1d2d881-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.167731 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mlfn\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-kube-api-access-4mlfn\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.167754 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/321fd323-3561-4655-b218-3972b1d2d881-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.167788 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.167814 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-bound-sa-token\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.199537 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.269358 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/321fd323-3561-4655-b218-3972b1d2d881-trusted-ca\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.269413 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/321fd323-3561-4655-b218-3972b1d2d881-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.269470 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mlfn\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-kube-api-access-4mlfn\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.269491 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/321fd323-3561-4655-b218-3972b1d2d881-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.269515 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-bound-sa-token\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.269557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/321fd323-3561-4655-b218-3972b1d2d881-registry-certificates\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.269579 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-registry-tls\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.270780 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/321fd323-3561-4655-b218-3972b1d2d881-trusted-ca\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.271035 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/321fd323-3561-4655-b218-3972b1d2d881-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.271957 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/321fd323-3561-4655-b218-3972b1d2d881-registry-certificates\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.276302 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/321fd323-3561-4655-b218-3972b1d2d881-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.285685 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mlfn\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-kube-api-access-4mlfn\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.285811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-bound-sa-token\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.291147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/321fd323-3561-4655-b218-3972b1d2d881-registry-tls\") pod \"image-registry-66df7c8f76-z49ph\" (UID: \"321fd323-3561-4655-b218-3972b1d2d881\") " pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.399812 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.577677 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z49ph"] Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.951438 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" event={"ID":"321fd323-3561-4655-b218-3972b1d2d881","Type":"ContainerStarted","Data":"ba577070c635c616def2a87610e61896f3416f47e9ce6989e36df29fe2bb1a6b"} Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.951500 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" event={"ID":"321fd323-3561-4655-b218-3972b1d2d881","Type":"ContainerStarted","Data":"1cf7a39ee7fb88677f982ceca3ded7cb815a8637929bf9c4890723403d8b10f4"} Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.952501 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:38 crc kubenswrapper[4726]: I1123 20:16:38.978055 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" podStartSLOduration=1.978035136 podStartE2EDuration="1.978035136s" podCreationTimestamp="2025-11-23 20:16:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:16:38.974512269 +0000 UTC m=+507.123553235" watchObservedRunningTime="2025-11-23 20:16:38.978035136 +0000 UTC m=+507.127076102" Nov 23 20:16:39 crc kubenswrapper[4726]: I1123 20:16:39.042771 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:16:39 crc kubenswrapper[4726]: I1123 20:16:39.043357 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:16:58 crc kubenswrapper[4726]: I1123 20:16:58.407678 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-z49ph" Nov 23 20:16:58 crc kubenswrapper[4726]: I1123 20:16:58.476456 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vg8gw"] Nov 23 20:17:09 crc kubenswrapper[4726]: I1123 20:17:09.042716 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:17:09 crc kubenswrapper[4726]: I1123 20:17:09.043610 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:17:23 crc kubenswrapper[4726]: I1123 20:17:23.580589 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" podUID="c928dd1a-1fe0-49d9-8829-608312d1de6f" containerName="registry" containerID="cri-o://ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306" gracePeriod=30 Nov 23 20:17:23 crc kubenswrapper[4726]: I1123 20:17:23.993238 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.098762 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-certificates\") pod \"c928dd1a-1fe0-49d9-8829-608312d1de6f\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.098821 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-tls\") pod \"c928dd1a-1fe0-49d9-8829-608312d1de6f\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.098918 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mxbk\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-kube-api-access-5mxbk\") pod \"c928dd1a-1fe0-49d9-8829-608312d1de6f\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.099051 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c928dd1a-1fe0-49d9-8829-608312d1de6f\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.099089 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-trusted-ca\") pod \"c928dd1a-1fe0-49d9-8829-608312d1de6f\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.099113 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c928dd1a-1fe0-49d9-8829-608312d1de6f-installation-pull-secrets\") pod \"c928dd1a-1fe0-49d9-8829-608312d1de6f\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.099151 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c928dd1a-1fe0-49d9-8829-608312d1de6f-ca-trust-extracted\") pod \"c928dd1a-1fe0-49d9-8829-608312d1de6f\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.099185 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-bound-sa-token\") pod \"c928dd1a-1fe0-49d9-8829-608312d1de6f\" (UID: \"c928dd1a-1fe0-49d9-8829-608312d1de6f\") " Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.100464 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c928dd1a-1fe0-49d9-8829-608312d1de6f" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.100484 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c928dd1a-1fe0-49d9-8829-608312d1de6f" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.106016 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c928dd1a-1fe0-49d9-8829-608312d1de6f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c928dd1a-1fe0-49d9-8829-608312d1de6f" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.106588 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-kube-api-access-5mxbk" (OuterVolumeSpecName: "kube-api-access-5mxbk") pod "c928dd1a-1fe0-49d9-8829-608312d1de6f" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f"). InnerVolumeSpecName "kube-api-access-5mxbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.108165 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c928dd1a-1fe0-49d9-8829-608312d1de6f" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.110357 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c928dd1a-1fe0-49d9-8829-608312d1de6f" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.111813 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c928dd1a-1fe0-49d9-8829-608312d1de6f" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.121654 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c928dd1a-1fe0-49d9-8829-608312d1de6f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c928dd1a-1fe0-49d9-8829-608312d1de6f" (UID: "c928dd1a-1fe0-49d9-8829-608312d1de6f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.200621 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.200653 4726 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.200671 4726 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.200682 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mxbk\" (UniqueName: \"kubernetes.io/projected/c928dd1a-1fe0-49d9-8829-608312d1de6f-kube-api-access-5mxbk\") on node \"crc\" DevicePath \"\"" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.200693 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c928dd1a-1fe0-49d9-8829-608312d1de6f-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.200704 4726 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c928dd1a-1fe0-49d9-8829-608312d1de6f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.200714 4726 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c928dd1a-1fe0-49d9-8829-608312d1de6f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.255891 4726 generic.go:334] "Generic (PLEG): container finished" podID="c928dd1a-1fe0-49d9-8829-608312d1de6f" containerID="ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306" exitCode=0 Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.255940 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" event={"ID":"c928dd1a-1fe0-49d9-8829-608312d1de6f","Type":"ContainerDied","Data":"ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306"} Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.255953 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.255971 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vg8gw" event={"ID":"c928dd1a-1fe0-49d9-8829-608312d1de6f","Type":"ContainerDied","Data":"074a3f85e8ac95e22d1c0ab468ae8bca246a1fe3036c18c56d93dcda4c24c6e8"} Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.255993 4726 scope.go:117] "RemoveContainer" containerID="ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.279923 4726 scope.go:117] "RemoveContainer" containerID="ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306" Nov 23 20:17:24 crc kubenswrapper[4726]: E1123 20:17:24.280428 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306\": container with ID starting with ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306 not found: ID does not exist" containerID="ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.280466 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306"} err="failed to get container status \"ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306\": rpc error: code = NotFound desc = could not find container \"ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306\": container with ID starting with ac2c7dddefbf8fca964c685a271f6af7e7ec99feed1969480dfefe5425eff306 not found: ID does not exist" Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.293864 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vg8gw"] Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.301293 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vg8gw"] Nov 23 20:17:24 crc kubenswrapper[4726]: I1123 20:17:24.594679 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c928dd1a-1fe0-49d9-8829-608312d1de6f" path="/var/lib/kubelet/pods/c928dd1a-1fe0-49d9-8829-608312d1de6f/volumes" Nov 23 20:17:39 crc kubenswrapper[4726]: I1123 20:17:39.043708 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:17:39 crc kubenswrapper[4726]: I1123 20:17:39.044987 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:17:39 crc kubenswrapper[4726]: I1123 20:17:39.045066 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:17:39 crc kubenswrapper[4726]: I1123 20:17:39.045936 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b42f4b47c86a8ff3e157d6e822c8e9bbf78256d76d3c430faf21561b98b7270"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:17:39 crc kubenswrapper[4726]: I1123 20:17:39.046029 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://5b42f4b47c86a8ff3e157d6e822c8e9bbf78256d76d3c430faf21561b98b7270" gracePeriod=600 Nov 23 20:17:39 crc kubenswrapper[4726]: I1123 20:17:39.384100 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="5b42f4b47c86a8ff3e157d6e822c8e9bbf78256d76d3c430faf21561b98b7270" exitCode=0 Nov 23 20:17:39 crc kubenswrapper[4726]: I1123 20:17:39.384175 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"5b42f4b47c86a8ff3e157d6e822c8e9bbf78256d76d3c430faf21561b98b7270"} Nov 23 20:17:39 crc kubenswrapper[4726]: I1123 20:17:39.384701 4726 scope.go:117] "RemoveContainer" containerID="d69e63c1f53e98006b9aa5285330cc02cef0dd0c90e397e35c78c1a1ac967d6f" Nov 23 20:17:40 crc kubenswrapper[4726]: I1123 20:17:40.398238 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"8ef166c526ceeb4bedeeddb339c120f70a8a458484a2d19fd740d49251ee0d41"} Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.548801 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dcnpn"] Nov 23 20:17:54 crc kubenswrapper[4726]: E1123 20:17:54.557540 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c928dd1a-1fe0-49d9-8829-608312d1de6f" containerName="registry" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.557650 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c928dd1a-1fe0-49d9-8829-608312d1de6f" containerName="registry" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.557881 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c928dd1a-1fe0-49d9-8829-608312d1de6f" containerName="registry" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.558412 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-dcnpn" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.566553 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-dcbkk" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.568855 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.569033 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.571406 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vp4hq"] Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.572133 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vp4hq" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.575524 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-b55jw" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.581298 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vp4hq"] Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.587498 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dcnpn"] Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.601710 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-llcjm"] Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.602321 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.607153 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-j6jwq" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.611550 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-llcjm"] Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.670378 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxksp\" (UniqueName: \"kubernetes.io/projected/b67043f1-eb7f-4239-b3a2-1f328e5f681f-kube-api-access-gxksp\") pod \"cert-manager-cainjector-7f985d654d-dcnpn\" (UID: \"b67043f1-eb7f-4239-b3a2-1f328e5f681f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dcnpn" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.670453 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgwxf\" (UniqueName: \"kubernetes.io/projected/d7287208-3876-4ffe-b893-a588d7383fe8-kube-api-access-wgwxf\") pod \"cert-manager-webhook-5655c58dd6-llcjm\" (UID: \"d7287208-3876-4ffe-b893-a588d7383fe8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.670548 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nklzp\" (UniqueName: \"kubernetes.io/projected/3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd-kube-api-access-nklzp\") pod \"cert-manager-5b446d88c5-vp4hq\" (UID: \"3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd\") " pod="cert-manager/cert-manager-5b446d88c5-vp4hq" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.771734 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxksp\" (UniqueName: \"kubernetes.io/projected/b67043f1-eb7f-4239-b3a2-1f328e5f681f-kube-api-access-gxksp\") pod \"cert-manager-cainjector-7f985d654d-dcnpn\" (UID: \"b67043f1-eb7f-4239-b3a2-1f328e5f681f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dcnpn" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.771782 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgwxf\" (UniqueName: \"kubernetes.io/projected/d7287208-3876-4ffe-b893-a588d7383fe8-kube-api-access-wgwxf\") pod \"cert-manager-webhook-5655c58dd6-llcjm\" (UID: \"d7287208-3876-4ffe-b893-a588d7383fe8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.771808 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nklzp\" (UniqueName: \"kubernetes.io/projected/3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd-kube-api-access-nklzp\") pod \"cert-manager-5b446d88c5-vp4hq\" (UID: \"3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd\") " pod="cert-manager/cert-manager-5b446d88c5-vp4hq" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.795187 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgwxf\" (UniqueName: \"kubernetes.io/projected/d7287208-3876-4ffe-b893-a588d7383fe8-kube-api-access-wgwxf\") pod \"cert-manager-webhook-5655c58dd6-llcjm\" (UID: \"d7287208-3876-4ffe-b893-a588d7383fe8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.797113 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxksp\" (UniqueName: \"kubernetes.io/projected/b67043f1-eb7f-4239-b3a2-1f328e5f681f-kube-api-access-gxksp\") pod \"cert-manager-cainjector-7f985d654d-dcnpn\" (UID: \"b67043f1-eb7f-4239-b3a2-1f328e5f681f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dcnpn" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.799238 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nklzp\" (UniqueName: \"kubernetes.io/projected/3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd-kube-api-access-nklzp\") pod \"cert-manager-5b446d88c5-vp4hq\" (UID: \"3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd\") " pod="cert-manager/cert-manager-5b446d88c5-vp4hq" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.890455 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-dcnpn" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.900074 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vp4hq" Nov 23 20:17:54 crc kubenswrapper[4726]: I1123 20:17:54.922254 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" Nov 23 20:17:55 crc kubenswrapper[4726]: I1123 20:17:55.136778 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dcnpn"] Nov 23 20:17:55 crc kubenswrapper[4726]: I1123 20:17:55.148081 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 20:17:55 crc kubenswrapper[4726]: I1123 20:17:55.445546 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-llcjm"] Nov 23 20:17:55 crc kubenswrapper[4726]: I1123 20:17:55.452577 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vp4hq"] Nov 23 20:17:55 crc kubenswrapper[4726]: I1123 20:17:55.495630 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-dcnpn" event={"ID":"b67043f1-eb7f-4239-b3a2-1f328e5f681f","Type":"ContainerStarted","Data":"b28f3fcdfc0c5694da06d8174a1e2b0cf5e97fe41fd584f6b722a530612b817c"} Nov 23 20:17:55 crc kubenswrapper[4726]: I1123 20:17:55.497415 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" event={"ID":"d7287208-3876-4ffe-b893-a588d7383fe8","Type":"ContainerStarted","Data":"2d2113cc11c118dffccc365e88eb67215b29baf26767b57d9da46f7ebabfb00c"} Nov 23 20:17:55 crc kubenswrapper[4726]: I1123 20:17:55.498925 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vp4hq" event={"ID":"3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd","Type":"ContainerStarted","Data":"aa49bbe7ad08c83595976226e1fa70ef2a0911561ce53d2dcd0d4c78a5aca0f4"} Nov 23 20:17:59 crc kubenswrapper[4726]: I1123 20:17:59.524353 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-dcnpn" event={"ID":"b67043f1-eb7f-4239-b3a2-1f328e5f681f","Type":"ContainerStarted","Data":"3fd21fca81369e7bb2a9dae8d36497fbb508dc5ccd83fba4d1096de2fb8055ce"} Nov 23 20:17:59 crc kubenswrapper[4726]: I1123 20:17:59.527322 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" event={"ID":"d7287208-3876-4ffe-b893-a588d7383fe8","Type":"ContainerStarted","Data":"2d087ebad7ea77bf635aa58fbcbd3ca8da80528bed95d399b3cd8a1b8b025762"} Nov 23 20:17:59 crc kubenswrapper[4726]: I1123 20:17:59.527663 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" Nov 23 20:17:59 crc kubenswrapper[4726]: I1123 20:17:59.529242 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vp4hq" event={"ID":"3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd","Type":"ContainerStarted","Data":"e7c1af10729842f0dd6e2096ad153a24c72e18eed8d7f9096bb21b9a03eb37b3"} Nov 23 20:17:59 crc kubenswrapper[4726]: I1123 20:17:59.543405 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-dcnpn" podStartSLOduration=1.967293262 podStartE2EDuration="5.543384555s" podCreationTimestamp="2025-11-23 20:17:54 +0000 UTC" firstStartedPulling="2025-11-23 20:17:55.147849069 +0000 UTC m=+583.296890025" lastFinishedPulling="2025-11-23 20:17:58.723940362 +0000 UTC m=+586.872981318" observedRunningTime="2025-11-23 20:17:59.541414847 +0000 UTC m=+587.690455803" watchObservedRunningTime="2025-11-23 20:17:59.543384555 +0000 UTC m=+587.692425511" Nov 23 20:17:59 crc kubenswrapper[4726]: I1123 20:17:59.559000 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-vp4hq" podStartSLOduration=2.305669627 podStartE2EDuration="5.558978523s" podCreationTimestamp="2025-11-23 20:17:54 +0000 UTC" firstStartedPulling="2025-11-23 20:17:55.469581457 +0000 UTC m=+583.618622413" lastFinishedPulling="2025-11-23 20:17:58.722890353 +0000 UTC m=+586.871931309" observedRunningTime="2025-11-23 20:17:59.55443842 +0000 UTC m=+587.703479396" watchObservedRunningTime="2025-11-23 20:17:59.558978523 +0000 UTC m=+587.708019479" Nov 23 20:17:59 crc kubenswrapper[4726]: I1123 20:17:59.572320 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" podStartSLOduration=2.325462635 podStartE2EDuration="5.572298594s" podCreationTimestamp="2025-11-23 20:17:54 +0000 UTC" firstStartedPulling="2025-11-23 20:17:55.467797929 +0000 UTC m=+583.616838915" lastFinishedPulling="2025-11-23 20:17:58.714633898 +0000 UTC m=+586.863674874" observedRunningTime="2025-11-23 20:17:59.569071884 +0000 UTC m=+587.718112860" watchObservedRunningTime="2025-11-23 20:17:59.572298594 +0000 UTC m=+587.721339540" Nov 23 20:18:04 crc kubenswrapper[4726]: I1123 20:18:04.927050 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-llcjm" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.055101 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-td7h8"] Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.055687 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovn-controller" containerID="cri-o://c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175" gracePeriod=30 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.055770 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="sbdb" containerID="cri-o://dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f" gracePeriod=30 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.055856 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kube-rbac-proxy-node" containerID="cri-o://26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae" gracePeriod=30 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.055811 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74" gracePeriod=30 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.055955 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovn-acl-logging" containerID="cri-o://8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6" gracePeriod=30 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.056024 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="nbdb" containerID="cri-o://ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94" gracePeriod=30 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.056402 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="northd" containerID="cri-o://402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128" gracePeriod=30 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.115332 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" containerID="cri-o://7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" gracePeriod=30 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.407439 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/3.log" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.409250 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovn-acl-logging/0.log" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.409734 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovn-controller/0.log" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.410118 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.459972 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jrc5t"] Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460199 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="nbdb" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460214 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="nbdb" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460231 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460240 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460252 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="sbdb" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460260 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="sbdb" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460270 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="northd" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460277 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="northd" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460286 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kube-rbac-proxy-node" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460294 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kube-rbac-proxy-node" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460306 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovn-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460313 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovn-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460324 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovn-acl-logging" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460333 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovn-acl-logging" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460348 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460357 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460365 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kubecfg-setup" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460376 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kubecfg-setup" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460385 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460393 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460402 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460409 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460420 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460427 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460539 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="sbdb" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460550 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="northd" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460561 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovn-acl-logging" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460571 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460579 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460589 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460597 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovn-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460609 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460619 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kube-rbac-proxy-node" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460626 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="nbdb" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460637 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="kube-rbac-proxy-ovn-metrics" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460646 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.460762 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.460772 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerName="ovnkube-controller" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.462786 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515122 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf84x\" (UniqueName: \"kubernetes.io/projected/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-kube-api-access-zf84x\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515175 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-script-lib\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515209 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-ovn\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515243 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-node-log\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515266 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-var-lib-openvswitch\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515286 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-netd\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515309 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-openvswitch\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515336 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-kubelet\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515370 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-var-lib-cni-networks-ovn-kubernetes\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515412 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-systemd\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-etc-openvswitch\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515472 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-netns\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515497 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-ovn-kubernetes\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515523 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-slash\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515547 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovn-node-metrics-cert\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515572 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-bin\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515597 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-systemd-units\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515621 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-log-socket\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515648 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-env-overrides\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515678 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-config\") pod \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\" (UID: \"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be\") " Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515815 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-node-log\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515841 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-env-overrides\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515867 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5pmr\" (UniqueName: \"kubernetes.io/projected/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-kube-api-access-p5pmr\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515913 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-slash\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515928 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515988 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-log-socket" (OuterVolumeSpecName: "log-socket") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516035 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516041 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516060 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516083 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516108 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516128 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-node-log" (OuterVolumeSpecName: "node-log") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516145 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516164 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516183 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-slash" (OuterVolumeSpecName: "host-slash") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516343 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516378 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516383 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.515944 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovnkube-script-lib\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516404 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516433 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-kubelet\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516433 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516470 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516562 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-etc-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516594 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovnkube-config\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516632 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovn-node-metrics-cert\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516651 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-cni-netd\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516673 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-run-ovn-kubernetes\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516779 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-systemd\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516818 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-cni-bin\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516874 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-var-lib-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516921 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-log-socket\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516947 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516970 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-run-netns\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.516986 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517009 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-systemd-units\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517029 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-ovn\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517088 4726 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517109 4726 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517118 4726 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517127 4726 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-slash\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517136 4726 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517144 4726 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517152 4726 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-log-socket\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517161 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517169 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517178 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517187 4726 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517207 4726 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-node-log\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517230 4726 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517247 4726 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517257 4726 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517267 4726 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.517279 4726 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.521337 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.521506 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-kube-api-access-zf84x" (OuterVolumeSpecName: "kube-api-access-zf84x") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "kube-api-access-zf84x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.527480 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" (UID: "db0ef36c-6b04-4a1f-bb8d-8f0895fd33be"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.568836 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovnkube-controller/3.log" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.570793 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovn-acl-logging/0.log" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571208 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-td7h8_db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/ovn-controller/0.log" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571521 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" exitCode=0 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571542 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f" exitCode=0 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571550 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94" exitCode=0 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571557 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128" exitCode=0 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571565 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74" exitCode=0 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571572 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae" exitCode=0 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571578 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6" exitCode=143 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571585 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" containerID="c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175" exitCode=143 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571585 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571625 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571638 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571649 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571659 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571668 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571673 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571679 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571794 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571639 4726 scope.go:117] "RemoveContainer" containerID="7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571802 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571917 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571928 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571934 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571940 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571946 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571951 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571966 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571982 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571988 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571993 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.571999 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572005 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572010 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572015 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572020 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572025 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572030 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572037 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572045 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572053 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572059 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572064 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572068 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572074 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572079 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572084 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572089 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572094 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572101 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-td7h8" event={"ID":"db0ef36c-6b04-4a1f-bb8d-8f0895fd33be","Type":"ContainerDied","Data":"504f88d9cf2135f28529f93c72c6e5ae0ca125dec7ee9de049b05ed7d5c29c7a"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572111 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572118 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572124 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572129 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572134 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572139 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572145 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572150 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572156 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.572173 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.574316 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/2.log" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.615733 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/1.log" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.615786 4726 generic.go:334] "Generic (PLEG): container finished" podID="8c5753d1-1307-4d28-b08f-3588c14ec1bd" containerID="57a5a173a024da59a8d4855dfecd282479d0835d0fcbcce8009f17c4cb39b643" exitCode=2 Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.615815 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jttr8" event={"ID":"8c5753d1-1307-4d28-b08f-3588c14ec1bd","Type":"ContainerDied","Data":"57a5a173a024da59a8d4855dfecd282479d0835d0fcbcce8009f17c4cb39b643"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.615839 4726 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3"} Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.616459 4726 scope.go:117] "RemoveContainer" containerID="57a5a173a024da59a8d4855dfecd282479d0835d0fcbcce8009f17c4cb39b643" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.616756 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jttr8_openshift-multus(8c5753d1-1307-4d28-b08f-3588c14ec1bd)\"" pod="openshift-multus/multus-jttr8" podUID="8c5753d1-1307-4d28-b08f-3588c14ec1bd" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620004 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-systemd-units\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620043 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-ovn\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-node-log\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620109 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-env-overrides\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620119 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-systemd-units\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620168 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-node-log\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620128 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5pmr\" (UniqueName: \"kubernetes.io/projected/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-kube-api-access-p5pmr\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620206 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-slash\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620213 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-ovn\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620231 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovnkube-script-lib\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620279 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-slash\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620312 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-kubelet\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620541 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-kubelet\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620660 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-etc-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620795 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-env-overrides\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.620866 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovnkube-script-lib\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.622051 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-etc-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.622078 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovnkube-config\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623057 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovn-node-metrics-cert\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623102 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-cni-netd\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623210 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovnkube-config\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623229 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-run-ovn-kubernetes\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623249 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-cni-netd\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623335 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-run-ovn-kubernetes\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623339 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-systemd\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623378 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-cni-bin\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623391 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-systemd\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623403 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-var-lib-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623430 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-log-socket\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623432 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-cni-bin\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623450 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-run-netns\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623487 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623497 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-log-socket\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623472 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-var-lib-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623632 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.623653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-host-run-netns\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.624188 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf84x\" (UniqueName: \"kubernetes.io/projected/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-kube-api-access-zf84x\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.624210 4726 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.624220 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.624285 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-run-openvswitch\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.626713 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-ovn-node-metrics-cert\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.647199 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5pmr\" (UniqueName: \"kubernetes.io/projected/d5c6b4c9-d723-4551-bf6c-e4b51aabb847-kube-api-access-p5pmr\") pod \"ovnkube-node-jrc5t\" (UID: \"d5c6b4c9-d723-4551-bf6c-e4b51aabb847\") " pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.649656 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.656581 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-td7h8"] Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.659027 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-td7h8"] Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.693516 4726 scope.go:117] "RemoveContainer" containerID="dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.711524 4726 scope.go:117] "RemoveContainer" containerID="ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.736391 4726 scope.go:117] "RemoveContainer" containerID="402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.747105 4726 scope.go:117] "RemoveContainer" containerID="957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.757579 4726 scope.go:117] "RemoveContainer" containerID="26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.767041 4726 scope.go:117] "RemoveContainer" containerID="8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.776665 4726 scope.go:117] "RemoveContainer" containerID="c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.777860 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.794230 4726 scope.go:117] "RemoveContainer" containerID="fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.805802 4726 scope.go:117] "RemoveContainer" containerID="7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.806267 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": container with ID starting with 7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa not found: ID does not exist" containerID="7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.806298 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} err="failed to get container status \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": rpc error: code = NotFound desc = could not find container \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": container with ID starting with 7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.806318 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.806524 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": container with ID starting with 85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92 not found: ID does not exist" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.806549 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} err="failed to get container status \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": rpc error: code = NotFound desc = could not find container \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": container with ID starting with 85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.806565 4726 scope.go:117] "RemoveContainer" containerID="dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.807157 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": container with ID starting with dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f not found: ID does not exist" containerID="dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.807180 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} err="failed to get container status \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": rpc error: code = NotFound desc = could not find container \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": container with ID starting with dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.807194 4726 scope.go:117] "RemoveContainer" containerID="ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.808574 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": container with ID starting with ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94 not found: ID does not exist" containerID="ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.808610 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} err="failed to get container status \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": rpc error: code = NotFound desc = could not find container \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": container with ID starting with ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.808634 4726 scope.go:117] "RemoveContainer" containerID="402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.808899 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": container with ID starting with 402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128 not found: ID does not exist" containerID="402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.808923 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} err="failed to get container status \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": rpc error: code = NotFound desc = could not find container \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": container with ID starting with 402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.808941 4726 scope.go:117] "RemoveContainer" containerID="957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.809248 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": container with ID starting with 957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74 not found: ID does not exist" containerID="957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.809296 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} err="failed to get container status \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": rpc error: code = NotFound desc = could not find container \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": container with ID starting with 957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.809312 4726 scope.go:117] "RemoveContainer" containerID="26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.809808 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": container with ID starting with 26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae not found: ID does not exist" containerID="26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.809835 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} err="failed to get container status \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": rpc error: code = NotFound desc = could not find container \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": container with ID starting with 26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.809854 4726 scope.go:117] "RemoveContainer" containerID="8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.810065 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": container with ID starting with 8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6 not found: ID does not exist" containerID="8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.810087 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} err="failed to get container status \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": rpc error: code = NotFound desc = could not find container \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": container with ID starting with 8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.810099 4726 scope.go:117] "RemoveContainer" containerID="c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.810299 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": container with ID starting with c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175 not found: ID does not exist" containerID="c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.810341 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} err="failed to get container status \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": rpc error: code = NotFound desc = could not find container \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": container with ID starting with c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.810371 4726 scope.go:117] "RemoveContainer" containerID="fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e" Nov 23 20:18:05 crc kubenswrapper[4726]: E1123 20:18:05.810615 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": container with ID starting with fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e not found: ID does not exist" containerID="fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.810642 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} err="failed to get container status \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": rpc error: code = NotFound desc = could not find container \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": container with ID starting with fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.810688 4726 scope.go:117] "RemoveContainer" containerID="7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811022 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} err="failed to get container status \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": rpc error: code = NotFound desc = could not find container \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": container with ID starting with 7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811042 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811191 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} err="failed to get container status \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": rpc error: code = NotFound desc = could not find container \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": container with ID starting with 85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811211 4726 scope.go:117] "RemoveContainer" containerID="dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811420 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} err="failed to get container status \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": rpc error: code = NotFound desc = could not find container \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": container with ID starting with dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811439 4726 scope.go:117] "RemoveContainer" containerID="ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811665 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} err="failed to get container status \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": rpc error: code = NotFound desc = could not find container \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": container with ID starting with ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811688 4726 scope.go:117] "RemoveContainer" containerID="402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811961 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} err="failed to get container status \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": rpc error: code = NotFound desc = could not find container \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": container with ID starting with 402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.811982 4726 scope.go:117] "RemoveContainer" containerID="957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.812173 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} err="failed to get container status \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": rpc error: code = NotFound desc = could not find container \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": container with ID starting with 957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.812201 4726 scope.go:117] "RemoveContainer" containerID="26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.812429 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} err="failed to get container status \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": rpc error: code = NotFound desc = could not find container \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": container with ID starting with 26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.812449 4726 scope.go:117] "RemoveContainer" containerID="8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.812621 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} err="failed to get container status \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": rpc error: code = NotFound desc = could not find container \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": container with ID starting with 8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.812641 4726 scope.go:117] "RemoveContainer" containerID="c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.812814 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} err="failed to get container status \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": rpc error: code = NotFound desc = could not find container \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": container with ID starting with c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.812838 4726 scope.go:117] "RemoveContainer" containerID="fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.813007 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} err="failed to get container status \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": rpc error: code = NotFound desc = could not find container \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": container with ID starting with fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.813024 4726 scope.go:117] "RemoveContainer" containerID="7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.813212 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} err="failed to get container status \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": rpc error: code = NotFound desc = could not find container \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": container with ID starting with 7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.813237 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.813419 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} err="failed to get container status \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": rpc error: code = NotFound desc = could not find container \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": container with ID starting with 85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.813442 4726 scope.go:117] "RemoveContainer" containerID="dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.813777 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} err="failed to get container status \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": rpc error: code = NotFound desc = could not find container \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": container with ID starting with dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.813798 4726 scope.go:117] "RemoveContainer" containerID="ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814025 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} err="failed to get container status \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": rpc error: code = NotFound desc = could not find container \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": container with ID starting with ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814045 4726 scope.go:117] "RemoveContainer" containerID="402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814363 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} err="failed to get container status \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": rpc error: code = NotFound desc = could not find container \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": container with ID starting with 402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814389 4726 scope.go:117] "RemoveContainer" containerID="957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814579 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} err="failed to get container status \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": rpc error: code = NotFound desc = could not find container \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": container with ID starting with 957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814596 4726 scope.go:117] "RemoveContainer" containerID="26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814771 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} err="failed to get container status \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": rpc error: code = NotFound desc = could not find container \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": container with ID starting with 26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814787 4726 scope.go:117] "RemoveContainer" containerID="8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814979 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} err="failed to get container status \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": rpc error: code = NotFound desc = could not find container \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": container with ID starting with 8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.814997 4726 scope.go:117] "RemoveContainer" containerID="c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.815194 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} err="failed to get container status \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": rpc error: code = NotFound desc = could not find container \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": container with ID starting with c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.815239 4726 scope.go:117] "RemoveContainer" containerID="fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.815445 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} err="failed to get container status \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": rpc error: code = NotFound desc = could not find container \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": container with ID starting with fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.815465 4726 scope.go:117] "RemoveContainer" containerID="7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.815676 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} err="failed to get container status \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": rpc error: code = NotFound desc = could not find container \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": container with ID starting with 7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.815694 4726 scope.go:117] "RemoveContainer" containerID="85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.816003 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92"} err="failed to get container status \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": rpc error: code = NotFound desc = could not find container \"85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92\": container with ID starting with 85a5cd68a7671ca7e820bf13e180dfea415f0709a514c548aa0f1335dc1fef92 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.816031 4726 scope.go:117] "RemoveContainer" containerID="dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.816337 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f"} err="failed to get container status \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": rpc error: code = NotFound desc = could not find container \"dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f\": container with ID starting with dda8accb1e99414fccf7bad561c3bd3bf360de4d4a5019bd2a77f60268a6942f not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.816359 4726 scope.go:117] "RemoveContainer" containerID="ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.816541 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94"} err="failed to get container status \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": rpc error: code = NotFound desc = could not find container \"ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94\": container with ID starting with ffd4172ee0a69c0660c338f5b0a6097f5d38950f60a03a171b2cb9f899290c94 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.816562 4726 scope.go:117] "RemoveContainer" containerID="402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.816797 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128"} err="failed to get container status \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": rpc error: code = NotFound desc = could not find container \"402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128\": container with ID starting with 402a2023e4b0258fc5d0bb7e36d1cd430bfb879c9937667abdb3a95a44c3b128 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.816818 4726 scope.go:117] "RemoveContainer" containerID="957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817056 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74"} err="failed to get container status \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": rpc error: code = NotFound desc = could not find container \"957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74\": container with ID starting with 957f5697b2f2fe7e7785edb7d545c654ac479c3da4bc2176b930927543581c74 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817075 4726 scope.go:117] "RemoveContainer" containerID="26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817256 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae"} err="failed to get container status \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": rpc error: code = NotFound desc = could not find container \"26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae\": container with ID starting with 26a756532b0999032b606136b46df773c59a0a4ba2e0855253ff241bb739f4ae not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817275 4726 scope.go:117] "RemoveContainer" containerID="8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817472 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6"} err="failed to get container status \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": rpc error: code = NotFound desc = could not find container \"8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6\": container with ID starting with 8518aeeddcd97f18d7e255439bb30974460cd455bc1c45a6ec798b9488bceca6 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817533 4726 scope.go:117] "RemoveContainer" containerID="c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817724 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175"} err="failed to get container status \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": rpc error: code = NotFound desc = could not find container \"c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175\": container with ID starting with c45e681d9f4878067606f4ce4c912347a2cd3f386419f037a87846af9cfa3175 not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817743 4726 scope.go:117] "RemoveContainer" containerID="fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817932 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e"} err="failed to get container status \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": rpc error: code = NotFound desc = could not find container \"fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e\": container with ID starting with fdc8d741abd0819dd8d0eb362bf995ee1e4670046ab16f3a391c398364979f1e not found: ID does not exist" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.817951 4726 scope.go:117] "RemoveContainer" containerID="7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa" Nov 23 20:18:05 crc kubenswrapper[4726]: I1123 20:18:05.818138 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa"} err="failed to get container status \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": rpc error: code = NotFound desc = could not find container \"7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa\": container with ID starting with 7b658b6422df99e72413964266f0b1052cbd8e71a3d5a2ed72d5911dd06732aa not found: ID does not exist" Nov 23 20:18:06 crc kubenswrapper[4726]: I1123 20:18:06.597671 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db0ef36c-6b04-4a1f-bb8d-8f0895fd33be" path="/var/lib/kubelet/pods/db0ef36c-6b04-4a1f-bb8d-8f0895fd33be/volumes" Nov 23 20:18:06 crc kubenswrapper[4726]: I1123 20:18:06.625314 4726 generic.go:334] "Generic (PLEG): container finished" podID="d5c6b4c9-d723-4551-bf6c-e4b51aabb847" containerID="cd1b1d4fcf86876e0f923ec1d710a56f1520dcb7d193408fb7365075c231c6d3" exitCode=0 Nov 23 20:18:06 crc kubenswrapper[4726]: I1123 20:18:06.625380 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerDied","Data":"cd1b1d4fcf86876e0f923ec1d710a56f1520dcb7d193408fb7365075c231c6d3"} Nov 23 20:18:06 crc kubenswrapper[4726]: I1123 20:18:06.625420 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"dabeb89759ec7cd1876a1cb661901655b15eaa292a471670c68eb5f0cafc0343"} Nov 23 20:18:07 crc kubenswrapper[4726]: I1123 20:18:07.631931 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"a9d9cfa298306d9603ec0a8fa457e77885ff5cc3da5b44f72879e79247313dee"} Nov 23 20:18:07 crc kubenswrapper[4726]: I1123 20:18:07.632215 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"444d934755659358e5ca0fe64713ca2cc98160e9d48d95cdec2b14dffc5eb705"} Nov 23 20:18:07 crc kubenswrapper[4726]: I1123 20:18:07.632234 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"374d834e701bac506c3595b575d930374c444221492cff32c1c918a73c9cd26a"} Nov 23 20:18:07 crc kubenswrapper[4726]: I1123 20:18:07.632246 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"729aea1ca0b38ce88ee587ba39b4c7bfc8c4bbc5f0fd19425836e7a3da1238cc"} Nov 23 20:18:07 crc kubenswrapper[4726]: I1123 20:18:07.632257 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"69e5344a0b01b41a43e0479eec4bdd789f5e7d0ff4d921dba774354bf89723ef"} Nov 23 20:18:07 crc kubenswrapper[4726]: I1123 20:18:07.632267 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"97cc3600266fb92d05c1e5936b331bfd9ecf6891bf781082b15c97e63ed123c4"} Nov 23 20:18:10 crc kubenswrapper[4726]: I1123 20:18:10.658588 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"649e06fd1470beca4ef66b37d0228c9766c6aa509faf12944afe3ae14c5ef0c7"} Nov 23 20:18:12 crc kubenswrapper[4726]: I1123 20:18:12.671255 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" event={"ID":"d5c6b4c9-d723-4551-bf6c-e4b51aabb847","Type":"ContainerStarted","Data":"09a2d7896056268e6bc9f9344cdbcb7d09032c7e81144e8d2bfc44baff6f25b6"} Nov 23 20:18:12 crc kubenswrapper[4726]: I1123 20:18:12.672496 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:12 crc kubenswrapper[4726]: I1123 20:18:12.672532 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:12 crc kubenswrapper[4726]: I1123 20:18:12.672586 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:12 crc kubenswrapper[4726]: I1123 20:18:12.703794 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:12 crc kubenswrapper[4726]: I1123 20:18:12.708858 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" podStartSLOduration=7.708842778 podStartE2EDuration="7.708842778s" podCreationTimestamp="2025-11-23 20:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:18:12.705637858 +0000 UTC m=+600.854678844" watchObservedRunningTime="2025-11-23 20:18:12.708842778 +0000 UTC m=+600.857883734" Nov 23 20:18:12 crc kubenswrapper[4726]: I1123 20:18:12.716418 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:12 crc kubenswrapper[4726]: I1123 20:18:12.763176 4726 scope.go:117] "RemoveContainer" containerID="268779a69aec6d27d0f40ab952fad0114867f9bf195a4410cf6edeec33be3ba3" Nov 23 20:18:13 crc kubenswrapper[4726]: I1123 20:18:13.681936 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/2.log" Nov 23 20:18:20 crc kubenswrapper[4726]: I1123 20:18:20.588746 4726 scope.go:117] "RemoveContainer" containerID="57a5a173a024da59a8d4855dfecd282479d0835d0fcbcce8009f17c4cb39b643" Nov 23 20:18:20 crc kubenswrapper[4726]: E1123 20:18:20.589579 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jttr8_openshift-multus(8c5753d1-1307-4d28-b08f-3588c14ec1bd)\"" pod="openshift-multus/multus-jttr8" podUID="8c5753d1-1307-4d28-b08f-3588c14ec1bd" Nov 23 20:18:34 crc kubenswrapper[4726]: I1123 20:18:34.589314 4726 scope.go:117] "RemoveContainer" containerID="57a5a173a024da59a8d4855dfecd282479d0835d0fcbcce8009f17c4cb39b643" Nov 23 20:18:34 crc kubenswrapper[4726]: I1123 20:18:34.819044 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jttr8_8c5753d1-1307-4d28-b08f-3588c14ec1bd/kube-multus/2.log" Nov 23 20:18:34 crc kubenswrapper[4726]: I1123 20:18:34.819427 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jttr8" event={"ID":"8c5753d1-1307-4d28-b08f-3588c14ec1bd","Type":"ContainerStarted","Data":"e3a79490224da21d18e18d1f755e2943d6164c7ca3a9337033c3817d89c79d76"} Nov 23 20:18:35 crc kubenswrapper[4726]: I1123 20:18:35.853614 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jrc5t" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.609727 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h"] Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.611313 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.613431 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.635827 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h"] Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.683148 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.683265 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnzbr\" (UniqueName: \"kubernetes.io/projected/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-kube-api-access-pnzbr\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.683323 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.784299 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.784372 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnzbr\" (UniqueName: \"kubernetes.io/projected/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-kube-api-access-pnzbr\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.784404 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.784843 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.785695 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.812554 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnzbr\" (UniqueName: \"kubernetes.io/projected/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-kube-api-access-pnzbr\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:46 crc kubenswrapper[4726]: I1123 20:18:46.926238 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:47 crc kubenswrapper[4726]: I1123 20:18:47.242529 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h"] Nov 23 20:18:47 crc kubenswrapper[4726]: W1123 20:18:47.276204 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71acdd00_d9b5_4b6a_a454_f8f65c4a5365.slice/crio-57b6b4ba1bb1a6298d0a3ea2ee2a14db2a1a17d2fe08625d6cf1b8e180fa844d WatchSource:0}: Error finding container 57b6b4ba1bb1a6298d0a3ea2ee2a14db2a1a17d2fe08625d6cf1b8e180fa844d: Status 404 returned error can't find the container with id 57b6b4ba1bb1a6298d0a3ea2ee2a14db2a1a17d2fe08625d6cf1b8e180fa844d Nov 23 20:18:47 crc kubenswrapper[4726]: I1123 20:18:47.905484 4726 generic.go:334] "Generic (PLEG): container finished" podID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerID="b473a2c86f1ea7ab692b7bd1166fbdb7621c01deaff8dc786ab9ed0ceee1b440" exitCode=0 Nov 23 20:18:47 crc kubenswrapper[4726]: I1123 20:18:47.905642 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" event={"ID":"71acdd00-d9b5-4b6a-a454-f8f65c4a5365","Type":"ContainerDied","Data":"b473a2c86f1ea7ab692b7bd1166fbdb7621c01deaff8dc786ab9ed0ceee1b440"} Nov 23 20:18:47 crc kubenswrapper[4726]: I1123 20:18:47.905852 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" event={"ID":"71acdd00-d9b5-4b6a-a454-f8f65c4a5365","Type":"ContainerStarted","Data":"57b6b4ba1bb1a6298d0a3ea2ee2a14db2a1a17d2fe08625d6cf1b8e180fa844d"} Nov 23 20:18:49 crc kubenswrapper[4726]: I1123 20:18:49.918701 4726 generic.go:334] "Generic (PLEG): container finished" podID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerID="d3b3804aa949ef6df5d2361836c934df6789b33c32ecdb978179f39bceeaff33" exitCode=0 Nov 23 20:18:49 crc kubenswrapper[4726]: I1123 20:18:49.918783 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" event={"ID":"71acdd00-d9b5-4b6a-a454-f8f65c4a5365","Type":"ContainerDied","Data":"d3b3804aa949ef6df5d2361836c934df6789b33c32ecdb978179f39bceeaff33"} Nov 23 20:18:50 crc kubenswrapper[4726]: I1123 20:18:50.967980 4726 generic.go:334] "Generic (PLEG): container finished" podID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerID="76b5b81cd048f1fc57e8eb623796d95e5c92686d4a48ce6ee4332cd135ad8f32" exitCode=0 Nov 23 20:18:50 crc kubenswrapper[4726]: I1123 20:18:50.968060 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" event={"ID":"71acdd00-d9b5-4b6a-a454-f8f65c4a5365","Type":"ContainerDied","Data":"76b5b81cd048f1fc57e8eb623796d95e5c92686d4a48ce6ee4332cd135ad8f32"} Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.327742 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.453535 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-bundle\") pod \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.453637 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnzbr\" (UniqueName: \"kubernetes.io/projected/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-kube-api-access-pnzbr\") pod \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.453719 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-util\") pod \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\" (UID: \"71acdd00-d9b5-4b6a-a454-f8f65c4a5365\") " Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.454697 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-bundle" (OuterVolumeSpecName: "bundle") pod "71acdd00-d9b5-4b6a-a454-f8f65c4a5365" (UID: "71acdd00-d9b5-4b6a-a454-f8f65c4a5365"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.455276 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.459552 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-kube-api-access-pnzbr" (OuterVolumeSpecName: "kube-api-access-pnzbr") pod "71acdd00-d9b5-4b6a-a454-f8f65c4a5365" (UID: "71acdd00-d9b5-4b6a-a454-f8f65c4a5365"). InnerVolumeSpecName "kube-api-access-pnzbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.558508 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnzbr\" (UniqueName: \"kubernetes.io/projected/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-kube-api-access-pnzbr\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.762756 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-util" (OuterVolumeSpecName: "util") pod "71acdd00-d9b5-4b6a-a454-f8f65c4a5365" (UID: "71acdd00-d9b5-4b6a-a454-f8f65c4a5365"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.861505 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71acdd00-d9b5-4b6a-a454-f8f65c4a5365-util\") on node \"crc\" DevicePath \"\"" Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.987019 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" event={"ID":"71acdd00-d9b5-4b6a-a454-f8f65c4a5365","Type":"ContainerDied","Data":"57b6b4ba1bb1a6298d0a3ea2ee2a14db2a1a17d2fe08625d6cf1b8e180fa844d"} Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.987294 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57b6b4ba1bb1a6298d0a3ea2ee2a14db2a1a17d2fe08625d6cf1b8e180fa844d" Nov 23 20:18:52 crc kubenswrapper[4726]: I1123 20:18:52.987139 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.267837 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-w8pws"] Nov 23 20:18:55 crc kubenswrapper[4726]: E1123 20:18:55.268026 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerName="extract" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.268038 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerName="extract" Nov 23 20:18:55 crc kubenswrapper[4726]: E1123 20:18:55.268054 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerName="pull" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.268060 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerName="pull" Nov 23 20:18:55 crc kubenswrapper[4726]: E1123 20:18:55.268071 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerName="util" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.268077 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerName="util" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.268155 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="71acdd00-d9b5-4b6a-a454-f8f65c4a5365" containerName="extract" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.268474 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-w8pws" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.270283 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-rkpxt" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.270527 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.270981 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.315233 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-w8pws"] Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.394163 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjnt5\" (UniqueName: \"kubernetes.io/projected/0debdf86-4fa6-46aa-8ffc-513dbb889153-kube-api-access-qjnt5\") pod \"nmstate-operator-557fdffb88-w8pws\" (UID: \"0debdf86-4fa6-46aa-8ffc-513dbb889153\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-w8pws" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.494964 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjnt5\" (UniqueName: \"kubernetes.io/projected/0debdf86-4fa6-46aa-8ffc-513dbb889153-kube-api-access-qjnt5\") pod \"nmstate-operator-557fdffb88-w8pws\" (UID: \"0debdf86-4fa6-46aa-8ffc-513dbb889153\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-w8pws" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.516938 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjnt5\" (UniqueName: \"kubernetes.io/projected/0debdf86-4fa6-46aa-8ffc-513dbb889153-kube-api-access-qjnt5\") pod \"nmstate-operator-557fdffb88-w8pws\" (UID: \"0debdf86-4fa6-46aa-8ffc-513dbb889153\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-w8pws" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.582169 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-w8pws" Nov 23 20:18:55 crc kubenswrapper[4726]: I1123 20:18:55.757001 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-w8pws"] Nov 23 20:18:56 crc kubenswrapper[4726]: I1123 20:18:56.000908 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-w8pws" event={"ID":"0debdf86-4fa6-46aa-8ffc-513dbb889153","Type":"ContainerStarted","Data":"1488e1b34cbe9e8366db1f423de188c2cb1f80bedd63be58621a16f0c375280a"} Nov 23 20:18:59 crc kubenswrapper[4726]: I1123 20:18:59.019321 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-w8pws" event={"ID":"0debdf86-4fa6-46aa-8ffc-513dbb889153","Type":"ContainerStarted","Data":"2f2b47b92b906c24e4b5722de5fdd307fe4de39e69c940d196f2efceeadf1e0f"} Nov 23 20:18:59 crc kubenswrapper[4726]: I1123 20:18:59.055961 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-w8pws" podStartSLOduration=1.7118330670000002 podStartE2EDuration="4.055924666s" podCreationTimestamp="2025-11-23 20:18:55 +0000 UTC" firstStartedPulling="2025-11-23 20:18:55.768620262 +0000 UTC m=+643.917661218" lastFinishedPulling="2025-11-23 20:18:58.112711861 +0000 UTC m=+646.261752817" observedRunningTime="2025-11-23 20:18:59.045929068 +0000 UTC m=+647.194970094" watchObservedRunningTime="2025-11-23 20:18:59.055924666 +0000 UTC m=+647.204965682" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.113113 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.113955 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.126476 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-qwj8c" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.130537 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.146960 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.147665 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.149598 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.155300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xj2l\" (UniqueName: \"kubernetes.io/projected/f89596fa-328e-483b-a5dd-34a241972f63-kube-api-access-6xj2l\") pod \"nmstate-webhook-6b89b748d8-jflhl\" (UID: \"f89596fa-328e-483b-a5dd-34a241972f63\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.155361 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f89596fa-328e-483b-a5dd-34a241972f63-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-jflhl\" (UID: \"f89596fa-328e-483b-a5dd-34a241972f63\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.155386 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5zzs\" (UniqueName: \"kubernetes.io/projected/fdc047b2-c091-49cc-8a3c-9663b20e7cd1-kube-api-access-q5zzs\") pod \"nmstate-metrics-5dcf9c57c5-pnnn6\" (UID: \"fdc047b2-c091-49cc-8a3c-9663b20e7cd1\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.159849 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.193028 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-w9sld"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.193645 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.256527 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f89596fa-328e-483b-a5dd-34a241972f63-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-jflhl\" (UID: \"f89596fa-328e-483b-a5dd-34a241972f63\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.256584 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5zzs\" (UniqueName: \"kubernetes.io/projected/fdc047b2-c091-49cc-8a3c-9663b20e7cd1-kube-api-access-q5zzs\") pod \"nmstate-metrics-5dcf9c57c5-pnnn6\" (UID: \"fdc047b2-c091-49cc-8a3c-9663b20e7cd1\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.256651 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xj2l\" (UniqueName: \"kubernetes.io/projected/f89596fa-328e-483b-a5dd-34a241972f63-kube-api-access-6xj2l\") pod \"nmstate-webhook-6b89b748d8-jflhl\" (UID: \"f89596fa-328e-483b-a5dd-34a241972f63\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:00 crc kubenswrapper[4726]: E1123 20:19:00.256940 4726 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 23 20:19:00 crc kubenswrapper[4726]: E1123 20:19:00.257035 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f89596fa-328e-483b-a5dd-34a241972f63-tls-key-pair podName:f89596fa-328e-483b-a5dd-34a241972f63 nodeName:}" failed. No retries permitted until 2025-11-23 20:19:00.757005793 +0000 UTC m=+648.906046809 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/f89596fa-328e-483b-a5dd-34a241972f63-tls-key-pair") pod "nmstate-webhook-6b89b748d8-jflhl" (UID: "f89596fa-328e-483b-a5dd-34a241972f63") : secret "openshift-nmstate-webhook" not found Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.279761 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.280383 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.282532 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-grn84" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.287596 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.287748 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.295909 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5zzs\" (UniqueName: \"kubernetes.io/projected/fdc047b2-c091-49cc-8a3c-9663b20e7cd1-kube-api-access-q5zzs\") pod \"nmstate-metrics-5dcf9c57c5-pnnn6\" (UID: \"fdc047b2-c091-49cc-8a3c-9663b20e7cd1\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.299476 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xj2l\" (UniqueName: \"kubernetes.io/projected/f89596fa-328e-483b-a5dd-34a241972f63-kube-api-access-6xj2l\") pod \"nmstate-webhook-6b89b748d8-jflhl\" (UID: \"f89596fa-328e-483b-a5dd-34a241972f63\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.313219 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.358089 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-ovs-socket\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.358143 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jpsf\" (UniqueName: \"kubernetes.io/projected/97c85373-c730-40fd-8b73-0b163c59e6ba-kube-api-access-8jpsf\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.358406 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-dbus-socket\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.358486 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-nmstate-lock\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.451514 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.459751 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-nmstate-lock\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.459806 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh95q\" (UniqueName: \"kubernetes.io/projected/6dc6c946-84ef-4ba1-8753-efd769c877c0-kube-api-access-dh95q\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.459843 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-ovs-socket\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.459894 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6dc6c946-84ef-4ba1-8753-efd769c877c0-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.459896 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-nmstate-lock\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.459913 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jpsf\" (UniqueName: \"kubernetes.io/projected/97c85373-c730-40fd-8b73-0b163c59e6ba-kube-api-access-8jpsf\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.460691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-ovs-socket\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.460820 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-dbus-socket\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.460883 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6dc6c946-84ef-4ba1-8753-efd769c877c0-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.461351 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/97c85373-c730-40fd-8b73-0b163c59e6ba-dbus-socket\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.463618 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6bd8c7cd6d-dbsb7"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.464313 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.473256 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6bd8c7cd6d-dbsb7"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.496763 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jpsf\" (UniqueName: \"kubernetes.io/projected/97c85373-c730-40fd-8b73-0b163c59e6ba-kube-api-access-8jpsf\") pod \"nmstate-handler-w9sld\" (UID: \"97c85373-c730-40fd-8b73-0b163c59e6ba\") " pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.510296 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.561548 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-oauth-config\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.561619 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-serving-cert\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.561680 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6dc6c946-84ef-4ba1-8753-efd769c877c0-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.561744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh95q\" (UniqueName: \"kubernetes.io/projected/6dc6c946-84ef-4ba1-8753-efd769c877c0-kube-api-access-dh95q\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.561769 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-trusted-ca-bundle\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.561795 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-oauth-serving-cert\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.561821 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-service-ca\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.561963 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-config\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.562012 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsssh\" (UniqueName: \"kubernetes.io/projected/a4630c47-5007-4f9c-a595-b4c7ce17e368-kube-api-access-bsssh\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.562055 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6dc6c946-84ef-4ba1-8753-efd769c877c0-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.563726 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6dc6c946-84ef-4ba1-8753-efd769c877c0-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.578658 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6dc6c946-84ef-4ba1-8753-efd769c877c0-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.581608 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh95q\" (UniqueName: \"kubernetes.io/projected/6dc6c946-84ef-4ba1-8753-efd769c877c0-kube-api-access-dh95q\") pod \"nmstate-console-plugin-5874bd7bc5-98ljz\" (UID: \"6dc6c946-84ef-4ba1-8753-efd769c877c0\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.627055 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.662500 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-oauth-config\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.662529 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-serving-cert\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.662566 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-trusted-ca-bundle\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.662582 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-oauth-serving-cert\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.662595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-service-ca\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.662614 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-config\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.662633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsssh\" (UniqueName: \"kubernetes.io/projected/a4630c47-5007-4f9c-a595-b4c7ce17e368-kube-api-access-bsssh\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.664045 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-oauth-serving-cert\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.665401 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-service-ca\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.665579 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-trusted-ca-bundle\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.666313 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-config\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.666795 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-oauth-config\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.667208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4630c47-5007-4f9c-a595-b4c7ce17e368-console-serving-cert\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.682515 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsssh\" (UniqueName: \"kubernetes.io/projected/a4630c47-5007-4f9c-a595-b4c7ce17e368-kube-api-access-bsssh\") pod \"console-6bd8c7cd6d-dbsb7\" (UID: \"a4630c47-5007-4f9c-a595-b4c7ce17e368\") " pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.763335 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f89596fa-328e-483b-a5dd-34a241972f63-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-jflhl\" (UID: \"f89596fa-328e-483b-a5dd-34a241972f63\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.767663 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f89596fa-328e-483b-a5dd-34a241972f63-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-jflhl\" (UID: \"f89596fa-328e-483b-a5dd-34a241972f63\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.830479 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.858381 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz"] Nov 23 20:19:00 crc kubenswrapper[4726]: I1123 20:19:00.909355 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6"] Nov 23 20:19:00 crc kubenswrapper[4726]: W1123 20:19:00.927101 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdc047b2_c091_49cc_8a3c_9663b20e7cd1.slice/crio-221db918f1f752709bc9e12adea3ea523076aecaad3bba1c92eed50fe1daa2aa WatchSource:0}: Error finding container 221db918f1f752709bc9e12adea3ea523076aecaad3bba1c92eed50fe1daa2aa: Status 404 returned error can't find the container with id 221db918f1f752709bc9e12adea3ea523076aecaad3bba1c92eed50fe1daa2aa Nov 23 20:19:01 crc kubenswrapper[4726]: I1123 20:19:01.029080 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-w9sld" event={"ID":"97c85373-c730-40fd-8b73-0b163c59e6ba","Type":"ContainerStarted","Data":"f77afc24c114956993efed973ee7ad34f7b9614b9bbe23ac86b8852b9b17cd4e"} Nov 23 20:19:01 crc kubenswrapper[4726]: I1123 20:19:01.030202 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" event={"ID":"6dc6c946-84ef-4ba1-8753-efd769c877c0","Type":"ContainerStarted","Data":"630e62d82e1c29c0bb7a51cfedb907a47adcaacb846bf3074b6d2ff2496f9551"} Nov 23 20:19:01 crc kubenswrapper[4726]: I1123 20:19:01.030934 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" event={"ID":"fdc047b2-c091-49cc-8a3c-9663b20e7cd1","Type":"ContainerStarted","Data":"221db918f1f752709bc9e12adea3ea523076aecaad3bba1c92eed50fe1daa2aa"} Nov 23 20:19:01 crc kubenswrapper[4726]: I1123 20:19:01.053284 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6bd8c7cd6d-dbsb7"] Nov 23 20:19:01 crc kubenswrapper[4726]: I1123 20:19:01.065816 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:01 crc kubenswrapper[4726]: I1123 20:19:01.448607 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl"] Nov 23 20:19:01 crc kubenswrapper[4726]: W1123 20:19:01.457071 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf89596fa_328e_483b_a5dd_34a241972f63.slice/crio-e1bd6b6e0737dac1809d445bffc3c98e5bababa2d76895742782fd4f1570cc69 WatchSource:0}: Error finding container e1bd6b6e0737dac1809d445bffc3c98e5bababa2d76895742782fd4f1570cc69: Status 404 returned error can't find the container with id e1bd6b6e0737dac1809d445bffc3c98e5bababa2d76895742782fd4f1570cc69 Nov 23 20:19:02 crc kubenswrapper[4726]: I1123 20:19:02.037620 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6bd8c7cd6d-dbsb7" event={"ID":"a4630c47-5007-4f9c-a595-b4c7ce17e368","Type":"ContainerStarted","Data":"7d85147da5b37e4029d7b21bdf1fc4e45d1a1b9329220d6cc30c412768d00bbd"} Nov 23 20:19:02 crc kubenswrapper[4726]: I1123 20:19:02.037981 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6bd8c7cd6d-dbsb7" event={"ID":"a4630c47-5007-4f9c-a595-b4c7ce17e368","Type":"ContainerStarted","Data":"68d217b8c110fa0e122193e3baae607c7436c05e2bddc3ef75e149c486a688e8"} Nov 23 20:19:02 crc kubenswrapper[4726]: I1123 20:19:02.038896 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" event={"ID":"f89596fa-328e-483b-a5dd-34a241972f63","Type":"ContainerStarted","Data":"e1bd6b6e0737dac1809d445bffc3c98e5bababa2d76895742782fd4f1570cc69"} Nov 23 20:19:02 crc kubenswrapper[4726]: I1123 20:19:02.056623 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6bd8c7cd6d-dbsb7" podStartSLOduration=2.056585657 podStartE2EDuration="2.056585657s" podCreationTimestamp="2025-11-23 20:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:19:02.055775117 +0000 UTC m=+650.204816083" watchObservedRunningTime="2025-11-23 20:19:02.056585657 +0000 UTC m=+650.205626613" Nov 23 20:19:04 crc kubenswrapper[4726]: I1123 20:19:04.051842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" event={"ID":"f89596fa-328e-483b-a5dd-34a241972f63","Type":"ContainerStarted","Data":"b25a06fad59157a10ec49a3d38a2a3be08ce6be45cf0a47c51c1c3535fc92746"} Nov 23 20:19:04 crc kubenswrapper[4726]: I1123 20:19:04.052026 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:04 crc kubenswrapper[4726]: I1123 20:19:04.053443 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-w9sld" event={"ID":"97c85373-c730-40fd-8b73-0b163c59e6ba","Type":"ContainerStarted","Data":"5614b04e6097f920f75570b98f5bc1b81d46a15980f3dcfc306cdc025c0782e7"} Nov 23 20:19:04 crc kubenswrapper[4726]: I1123 20:19:04.053562 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:04 crc kubenswrapper[4726]: I1123 20:19:04.055118 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" event={"ID":"6dc6c946-84ef-4ba1-8753-efd769c877c0","Type":"ContainerStarted","Data":"f2373fc0ed7bfd30b8360e220bcefb10c1759d6d77f026cb2cb64b7306768444"} Nov 23 20:19:04 crc kubenswrapper[4726]: I1123 20:19:04.057072 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" event={"ID":"fdc047b2-c091-49cc-8a3c-9663b20e7cd1","Type":"ContainerStarted","Data":"6d4bdf15c2799272be26dd6cd975fb6790e8d4896baa0a5087dc952e60b66494"} Nov 23 20:19:04 crc kubenswrapper[4726]: I1123 20:19:04.081725 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" podStartSLOduration=2.17738353 podStartE2EDuration="4.081705947s" podCreationTimestamp="2025-11-23 20:19:00 +0000 UTC" firstStartedPulling="2025-11-23 20:19:01.459917925 +0000 UTC m=+649.608958881" lastFinishedPulling="2025-11-23 20:19:03.364240342 +0000 UTC m=+651.513281298" observedRunningTime="2025-11-23 20:19:04.076525208 +0000 UTC m=+652.225566174" watchObservedRunningTime="2025-11-23 20:19:04.081705947 +0000 UTC m=+652.230746903" Nov 23 20:19:04 crc kubenswrapper[4726]: I1123 20:19:04.097178 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-w9sld" podStartSLOduration=1.2678969310000001 podStartE2EDuration="4.097162671s" podCreationTimestamp="2025-11-23 20:19:00 +0000 UTC" firstStartedPulling="2025-11-23 20:19:00.531529797 +0000 UTC m=+648.680570753" lastFinishedPulling="2025-11-23 20:19:03.360795537 +0000 UTC m=+651.509836493" observedRunningTime="2025-11-23 20:19:04.089993423 +0000 UTC m=+652.239034409" watchObservedRunningTime="2025-11-23 20:19:04.097162671 +0000 UTC m=+652.246203627" Nov 23 20:19:06 crc kubenswrapper[4726]: I1123 20:19:06.071234 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" event={"ID":"fdc047b2-c091-49cc-8a3c-9663b20e7cd1","Type":"ContainerStarted","Data":"0609f6e4cec8af22d514681cdd47032863db30127e3932e0ed0ea66f6d930485"} Nov 23 20:19:06 crc kubenswrapper[4726]: I1123 20:19:06.090448 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-98ljz" podStartSLOduration=3.609464429 podStartE2EDuration="6.09042244s" podCreationTimestamp="2025-11-23 20:19:00 +0000 UTC" firstStartedPulling="2025-11-23 20:19:00.878680847 +0000 UTC m=+649.027721833" lastFinishedPulling="2025-11-23 20:19:03.359638888 +0000 UTC m=+651.508679844" observedRunningTime="2025-11-23 20:19:04.112240816 +0000 UTC m=+652.261281832" watchObservedRunningTime="2025-11-23 20:19:06.09042244 +0000 UTC m=+654.239463436" Nov 23 20:19:06 crc kubenswrapper[4726]: I1123 20:19:06.095027 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pnnn6" podStartSLOduration=1.547911732 podStartE2EDuration="6.095010681s" podCreationTimestamp="2025-11-23 20:19:00 +0000 UTC" firstStartedPulling="2025-11-23 20:19:00.928952007 +0000 UTC m=+649.077992963" lastFinishedPulling="2025-11-23 20:19:05.476050956 +0000 UTC m=+653.625091912" observedRunningTime="2025-11-23 20:19:06.090185773 +0000 UTC m=+654.239226749" watchObservedRunningTime="2025-11-23 20:19:06.095010681 +0000 UTC m=+654.244051677" Nov 23 20:19:10 crc kubenswrapper[4726]: I1123 20:19:10.546816 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-w9sld" Nov 23 20:19:10 crc kubenswrapper[4726]: I1123 20:19:10.831162 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:10 crc kubenswrapper[4726]: I1123 20:19:10.831667 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:10 crc kubenswrapper[4726]: I1123 20:19:10.838001 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:11 crc kubenswrapper[4726]: I1123 20:19:11.112783 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6bd8c7cd6d-dbsb7" Nov 23 20:19:11 crc kubenswrapper[4726]: I1123 20:19:11.174026 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-q4zwl"] Nov 23 20:19:21 crc kubenswrapper[4726]: I1123 20:19:21.071624 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-jflhl" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.019571 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx"] Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.022450 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.025490 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx"] Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.029770 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.129176 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.129242 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88t24\" (UniqueName: \"kubernetes.io/projected/ea8a3056-cbc8-4a11-9144-7f6a939e697d-kube-api-access-88t24\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.129322 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.230599 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.230675 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.230731 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88t24\" (UniqueName: \"kubernetes.io/projected/ea8a3056-cbc8-4a11-9144-7f6a939e697d-kube-api-access-88t24\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.231391 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.231689 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.252693 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88t24\" (UniqueName: \"kubernetes.io/projected/ea8a3056-cbc8-4a11-9144-7f6a939e697d-kube-api-access-88t24\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.353187 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:35 crc kubenswrapper[4726]: I1123 20:19:35.553087 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx"] Nov 23 20:19:36 crc kubenswrapper[4726]: I1123 20:19:36.230229 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-q4zwl" podUID="7445d8ee-ffb1-4610-ae2d-cb511edf1525" containerName="console" containerID="cri-o://6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c" gracePeriod=15 Nov 23 20:19:36 crc kubenswrapper[4726]: I1123 20:19:36.277368 4726 generic.go:334] "Generic (PLEG): container finished" podID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerID="7c5b2cedccec0925ddebe09f8d6908a2b4639c913189e699362ec9c643df6bb6" exitCode=0 Nov 23 20:19:36 crc kubenswrapper[4726]: I1123 20:19:36.277423 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" event={"ID":"ea8a3056-cbc8-4a11-9144-7f6a939e697d","Type":"ContainerDied","Data":"7c5b2cedccec0925ddebe09f8d6908a2b4639c913189e699362ec9c643df6bb6"} Nov 23 20:19:36 crc kubenswrapper[4726]: I1123 20:19:36.277473 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" event={"ID":"ea8a3056-cbc8-4a11-9144-7f6a939e697d","Type":"ContainerStarted","Data":"c73d8875ef9e17f8a6c0d237a706f491d3f8c122fad18e2b3e9fde621690f4eb"} Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.135750 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-q4zwl_7445d8ee-ffb1-4610-ae2d-cb511edf1525/console/0.log" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.136189 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.264312 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-service-ca\") pod \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.264362 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-config\") pod \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.264398 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td5r5\" (UniqueName: \"kubernetes.io/projected/7445d8ee-ffb1-4610-ae2d-cb511edf1525-kube-api-access-td5r5\") pod \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.264417 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-oauth-config\") pod \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.264444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-serving-cert\") pod \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.264471 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-oauth-serving-cert\") pod \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.264489 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-trusted-ca-bundle\") pod \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\" (UID: \"7445d8ee-ffb1-4610-ae2d-cb511edf1525\") " Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.265341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7445d8ee-ffb1-4610-ae2d-cb511edf1525" (UID: "7445d8ee-ffb1-4610-ae2d-cb511edf1525"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.266442 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-service-ca" (OuterVolumeSpecName: "service-ca") pod "7445d8ee-ffb1-4610-ae2d-cb511edf1525" (UID: "7445d8ee-ffb1-4610-ae2d-cb511edf1525"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.266792 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-config" (OuterVolumeSpecName: "console-config") pod "7445d8ee-ffb1-4610-ae2d-cb511edf1525" (UID: "7445d8ee-ffb1-4610-ae2d-cb511edf1525"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.267057 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7445d8ee-ffb1-4610-ae2d-cb511edf1525" (UID: "7445d8ee-ffb1-4610-ae2d-cb511edf1525"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.271722 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7445d8ee-ffb1-4610-ae2d-cb511edf1525-kube-api-access-td5r5" (OuterVolumeSpecName: "kube-api-access-td5r5") pod "7445d8ee-ffb1-4610-ae2d-cb511edf1525" (UID: "7445d8ee-ffb1-4610-ae2d-cb511edf1525"). InnerVolumeSpecName "kube-api-access-td5r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.272026 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7445d8ee-ffb1-4610-ae2d-cb511edf1525" (UID: "7445d8ee-ffb1-4610-ae2d-cb511edf1525"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.287370 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-q4zwl_7445d8ee-ffb1-4610-ae2d-cb511edf1525/console/0.log" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.287417 4726 generic.go:334] "Generic (PLEG): container finished" podID="7445d8ee-ffb1-4610-ae2d-cb511edf1525" containerID="6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c" exitCode=2 Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.287449 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-q4zwl" event={"ID":"7445d8ee-ffb1-4610-ae2d-cb511edf1525","Type":"ContainerDied","Data":"6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c"} Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.287477 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-q4zwl" event={"ID":"7445d8ee-ffb1-4610-ae2d-cb511edf1525","Type":"ContainerDied","Data":"8c440bd11882b91bdbf87e47af3c4de4ab04e2d0138649e1caa32f7d2922d045"} Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.287469 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7445d8ee-ffb1-4610-ae2d-cb511edf1525" (UID: "7445d8ee-ffb1-4610-ae2d-cb511edf1525"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.287497 4726 scope.go:117] "RemoveContainer" containerID="6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.287595 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-q4zwl" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.321780 4726 scope.go:117] "RemoveContainer" containerID="6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c" Nov 23 20:19:37 crc kubenswrapper[4726]: E1123 20:19:37.322649 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c\": container with ID starting with 6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c not found: ID does not exist" containerID="6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.322688 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c"} err="failed to get container status \"6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c\": rpc error: code = NotFound desc = could not find container \"6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c\": container with ID starting with 6f57bce8788797ea4c2619cc9d904505d6044a67ac9619c2f4a7a2443465370c not found: ID does not exist" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.326730 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-q4zwl"] Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.329936 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-q4zwl"] Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.365377 4726 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.365404 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td5r5\" (UniqueName: \"kubernetes.io/projected/7445d8ee-ffb1-4610-ae2d-cb511edf1525-kube-api-access-td5r5\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.365415 4726 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.365427 4726 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7445d8ee-ffb1-4610-ae2d-cb511edf1525-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.365438 4726 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.365449 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:37 crc kubenswrapper[4726]: I1123 20:19:37.365462 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7445d8ee-ffb1-4610-ae2d-cb511edf1525-service-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:38 crc kubenswrapper[4726]: I1123 20:19:38.294915 4726 generic.go:334] "Generic (PLEG): container finished" podID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerID="797604cfd82352bb6c2637275a783e0915f087da52a7da5ba3c3977410170743" exitCode=0 Nov 23 20:19:38 crc kubenswrapper[4726]: I1123 20:19:38.295030 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" event={"ID":"ea8a3056-cbc8-4a11-9144-7f6a939e697d","Type":"ContainerDied","Data":"797604cfd82352bb6c2637275a783e0915f087da52a7da5ba3c3977410170743"} Nov 23 20:19:38 crc kubenswrapper[4726]: I1123 20:19:38.598216 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7445d8ee-ffb1-4610-ae2d-cb511edf1525" path="/var/lib/kubelet/pods/7445d8ee-ffb1-4610-ae2d-cb511edf1525/volumes" Nov 23 20:19:39 crc kubenswrapper[4726]: I1123 20:19:39.043697 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:19:39 crc kubenswrapper[4726]: I1123 20:19:39.043787 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:19:39 crc kubenswrapper[4726]: I1123 20:19:39.305612 4726 generic.go:334] "Generic (PLEG): container finished" podID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerID="7036bb4d4b30d7037c108ccfb9e72c4d2b964a6d96b42b697401009bf7c86f22" exitCode=0 Nov 23 20:19:39 crc kubenswrapper[4726]: I1123 20:19:39.305671 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" event={"ID":"ea8a3056-cbc8-4a11-9144-7f6a939e697d","Type":"ContainerDied","Data":"7036bb4d4b30d7037c108ccfb9e72c4d2b964a6d96b42b697401009bf7c86f22"} Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.631275 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.807979 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-util\") pod \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.808045 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88t24\" (UniqueName: \"kubernetes.io/projected/ea8a3056-cbc8-4a11-9144-7f6a939e697d-kube-api-access-88t24\") pod \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.808205 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-bundle\") pod \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\" (UID: \"ea8a3056-cbc8-4a11-9144-7f6a939e697d\") " Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.809641 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-bundle" (OuterVolumeSpecName: "bundle") pod "ea8a3056-cbc8-4a11-9144-7f6a939e697d" (UID: "ea8a3056-cbc8-4a11-9144-7f6a939e697d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.817864 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea8a3056-cbc8-4a11-9144-7f6a939e697d-kube-api-access-88t24" (OuterVolumeSpecName: "kube-api-access-88t24") pod "ea8a3056-cbc8-4a11-9144-7f6a939e697d" (UID: "ea8a3056-cbc8-4a11-9144-7f6a939e697d"). InnerVolumeSpecName "kube-api-access-88t24". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.826596 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-util" (OuterVolumeSpecName: "util") pod "ea8a3056-cbc8-4a11-9144-7f6a939e697d" (UID: "ea8a3056-cbc8-4a11-9144-7f6a939e697d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.909899 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-util\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.909944 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88t24\" (UniqueName: \"kubernetes.io/projected/ea8a3056-cbc8-4a11-9144-7f6a939e697d-kube-api-access-88t24\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:40 crc kubenswrapper[4726]: I1123 20:19:40.909962 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea8a3056-cbc8-4a11-9144-7f6a939e697d-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:19:41 crc kubenswrapper[4726]: I1123 20:19:41.322135 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" event={"ID":"ea8a3056-cbc8-4a11-9144-7f6a939e697d","Type":"ContainerDied","Data":"c73d8875ef9e17f8a6c0d237a706f491d3f8c122fad18e2b3e9fde621690f4eb"} Nov 23 20:19:41 crc kubenswrapper[4726]: I1123 20:19:41.322479 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c73d8875ef9e17f8a6c0d237a706f491d3f8c122fad18e2b3e9fde621690f4eb" Nov 23 20:19:41 crc kubenswrapper[4726]: I1123 20:19:41.322542 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.154266 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t"] Nov 23 20:19:53 crc kubenswrapper[4726]: E1123 20:19:53.154967 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerName="util" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.154979 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerName="util" Nov 23 20:19:53 crc kubenswrapper[4726]: E1123 20:19:53.154989 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7445d8ee-ffb1-4610-ae2d-cb511edf1525" containerName="console" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.154996 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7445d8ee-ffb1-4610-ae2d-cb511edf1525" containerName="console" Nov 23 20:19:53 crc kubenswrapper[4726]: E1123 20:19:53.155010 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerName="extract" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.155016 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerName="extract" Nov 23 20:19:53 crc kubenswrapper[4726]: E1123 20:19:53.155024 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerName="pull" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.155029 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerName="pull" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.155145 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea8a3056-cbc8-4a11-9144-7f6a939e697d" containerName="extract" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.155176 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7445d8ee-ffb1-4610-ae2d-cb511edf1525" containerName="console" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.155598 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.160495 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.160665 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.160794 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fmdbv" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.160936 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.161234 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.177623 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t"] Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.297671 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be91600b-06c3-4acc-a0d6-d7e732a00228-apiservice-cert\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.297991 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be91600b-06c3-4acc-a0d6-d7e732a00228-webhook-cert\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.298125 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mlqf\" (UniqueName: \"kubernetes.io/projected/be91600b-06c3-4acc-a0d6-d7e732a00228-kube-api-access-2mlqf\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.400065 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be91600b-06c3-4acc-a0d6-d7e732a00228-webhook-cert\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.400104 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mlqf\" (UniqueName: \"kubernetes.io/projected/be91600b-06c3-4acc-a0d6-d7e732a00228-kube-api-access-2mlqf\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.400182 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be91600b-06c3-4acc-a0d6-d7e732a00228-apiservice-cert\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.406351 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be91600b-06c3-4acc-a0d6-d7e732a00228-webhook-cert\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.420402 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be91600b-06c3-4acc-a0d6-d7e732a00228-apiservice-cert\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.429598 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mlqf\" (UniqueName: \"kubernetes.io/projected/be91600b-06c3-4acc-a0d6-d7e732a00228-kube-api-access-2mlqf\") pod \"metallb-operator-controller-manager-64f7f896bc-tb44t\" (UID: \"be91600b-06c3-4acc-a0d6-d7e732a00228\") " pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.429777 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-c66969f79-8m676"] Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.430525 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.433434 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.433617 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-s4xnw" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.433732 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.470354 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.479932 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c66969f79-8m676"] Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.502954 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcwj2\" (UniqueName: \"kubernetes.io/projected/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-kube-api-access-bcwj2\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.502997 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-apiservice-cert\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.503019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-webhook-cert\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.605101 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcwj2\" (UniqueName: \"kubernetes.io/projected/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-kube-api-access-bcwj2\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.605624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-apiservice-cert\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.605645 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-webhook-cert\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.611962 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-webhook-cert\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.614409 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-apiservice-cert\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.628508 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcwj2\" (UniqueName: \"kubernetes.io/projected/bf3b7cb9-1dc8-4966-b15c-71d77aa2751b-kube-api-access-bcwj2\") pod \"metallb-operator-webhook-server-c66969f79-8m676\" (UID: \"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b\") " pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.812129 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:53 crc kubenswrapper[4726]: I1123 20:19:53.877741 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t"] Nov 23 20:19:53 crc kubenswrapper[4726]: W1123 20:19:53.925743 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe91600b_06c3_4acc_a0d6_d7e732a00228.slice/crio-f7bb0b6f70378d7d32fddafe3a25dafe84a42631ed9154a6a23c046dae2cbaaf WatchSource:0}: Error finding container f7bb0b6f70378d7d32fddafe3a25dafe84a42631ed9154a6a23c046dae2cbaaf: Status 404 returned error can't find the container with id f7bb0b6f70378d7d32fddafe3a25dafe84a42631ed9154a6a23c046dae2cbaaf Nov 23 20:19:54 crc kubenswrapper[4726]: I1123 20:19:54.159816 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c66969f79-8m676"] Nov 23 20:19:54 crc kubenswrapper[4726]: W1123 20:19:54.167100 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf3b7cb9_1dc8_4966_b15c_71d77aa2751b.slice/crio-b85f83ce51cc8fe71120e624265f3e2c81be5d8a3c6996ceb2e4dd4ed27e9736 WatchSource:0}: Error finding container b85f83ce51cc8fe71120e624265f3e2c81be5d8a3c6996ceb2e4dd4ed27e9736: Status 404 returned error can't find the container with id b85f83ce51cc8fe71120e624265f3e2c81be5d8a3c6996ceb2e4dd4ed27e9736 Nov 23 20:19:54 crc kubenswrapper[4726]: I1123 20:19:54.433259 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" event={"ID":"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b","Type":"ContainerStarted","Data":"b85f83ce51cc8fe71120e624265f3e2c81be5d8a3c6996ceb2e4dd4ed27e9736"} Nov 23 20:19:54 crc kubenswrapper[4726]: I1123 20:19:54.434134 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" event={"ID":"be91600b-06c3-4acc-a0d6-d7e732a00228","Type":"ContainerStarted","Data":"f7bb0b6f70378d7d32fddafe3a25dafe84a42631ed9154a6a23c046dae2cbaaf"} Nov 23 20:19:59 crc kubenswrapper[4726]: I1123 20:19:59.468068 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" event={"ID":"bf3b7cb9-1dc8-4966-b15c-71d77aa2751b","Type":"ContainerStarted","Data":"4eb945021ba1e88fabaa9a191a07930e60da42e9302fd5c97c553b0037a75464"} Nov 23 20:19:59 crc kubenswrapper[4726]: I1123 20:19:59.468685 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:19:59 crc kubenswrapper[4726]: I1123 20:19:59.469794 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" event={"ID":"be91600b-06c3-4acc-a0d6-d7e732a00228","Type":"ContainerStarted","Data":"eb705f1e678594f582c2e6aa2d56924fcfbec4fd36a7679867f1c0eafd1825a7"} Nov 23 20:19:59 crc kubenswrapper[4726]: I1123 20:19:59.469919 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:19:59 crc kubenswrapper[4726]: I1123 20:19:59.498888 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" podStartSLOduration=2.0837584749999998 podStartE2EDuration="6.498861241s" podCreationTimestamp="2025-11-23 20:19:53 +0000 UTC" firstStartedPulling="2025-11-23 20:19:54.169608695 +0000 UTC m=+702.318649651" lastFinishedPulling="2025-11-23 20:19:58.584711461 +0000 UTC m=+706.733752417" observedRunningTime="2025-11-23 20:19:59.492613362 +0000 UTC m=+707.641654318" watchObservedRunningTime="2025-11-23 20:19:59.498861241 +0000 UTC m=+707.647902197" Nov 23 20:19:59 crc kubenswrapper[4726]: I1123 20:19:59.523429 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" podStartSLOduration=1.9565285650000002 podStartE2EDuration="6.523411683s" podCreationTimestamp="2025-11-23 20:19:53 +0000 UTC" firstStartedPulling="2025-11-23 20:19:53.942441476 +0000 UTC m=+702.091482422" lastFinishedPulling="2025-11-23 20:19:58.509324584 +0000 UTC m=+706.658365540" observedRunningTime="2025-11-23 20:19:59.5233268 +0000 UTC m=+707.672367756" watchObservedRunningTime="2025-11-23 20:19:59.523411683 +0000 UTC m=+707.672452639" Nov 23 20:20:09 crc kubenswrapper[4726]: I1123 20:20:09.043100 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:20:09 crc kubenswrapper[4726]: I1123 20:20:09.043684 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:20:13 crc kubenswrapper[4726]: I1123 20:20:13.816558 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-c66969f79-8m676" Nov 23 20:20:33 crc kubenswrapper[4726]: I1123 20:20:33.474190 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-64f7f896bc-tb44t" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.234491 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp"] Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.235336 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.240760 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.241747 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-w9sjl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.249924 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp"] Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.258968 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-b8hdl"] Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.261036 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.269137 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.270546 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.360121 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-zkvn5"] Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.360899 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.369736 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-sockets\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.369761 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-2vb27"] Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.369793 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3130c09b-7002-4448-bdd4-87b73f70ff99-cert\") pod \"frr-k8s-webhook-server-6998585d5-mlnqp\" (UID: \"3130c09b-7002-4448-bdd4-87b73f70ff99\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.369814 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-reloader\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.369898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-conf\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.369926 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics-certs\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.370019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.370075 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xrmw\" (UniqueName: \"kubernetes.io/projected/3130c09b-7002-4448-bdd4-87b73f70ff99-kube-api-access-9xrmw\") pod \"frr-k8s-webhook-server-6998585d5-mlnqp\" (UID: \"3130c09b-7002-4448-bdd4-87b73f70ff99\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.370101 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-startup\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.370144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pz5w\" (UniqueName: \"kubernetes.io/projected/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-kube-api-access-2pz5w\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.371456 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.374038 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-9d227" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.374156 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.374171 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.374300 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.378084 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.392797 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-2vb27"] Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472010 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-cert\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472071 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-metrics-certs\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472088 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fnmb\" (UniqueName: \"kubernetes.io/projected/d99ab20e-6022-4d7e-8ff2-571cd0961a08-kube-api-access-9fnmb\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472122 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3130c09b-7002-4448-bdd4-87b73f70ff99-cert\") pod \"frr-k8s-webhook-server-6998585d5-mlnqp\" (UID: \"3130c09b-7002-4448-bdd4-87b73f70ff99\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472140 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-metrics-certs\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472156 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-reloader\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472180 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-conf\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472200 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d99ab20e-6022-4d7e-8ff2-571cd0961a08-metallb-excludel2\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472218 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics-certs\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472236 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472252 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472267 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xrmw\" (UniqueName: \"kubernetes.io/projected/3130c09b-7002-4448-bdd4-87b73f70ff99-kube-api-access-9xrmw\") pod \"frr-k8s-webhook-server-6998585d5-mlnqp\" (UID: \"3130c09b-7002-4448-bdd4-87b73f70ff99\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472283 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-startup\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pz5w\" (UniqueName: \"kubernetes.io/projected/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-kube-api-access-2pz5w\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472334 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l4s5\" (UniqueName: \"kubernetes.io/projected/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-kube-api-access-8l4s5\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472354 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-sockets\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.472838 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-sockets\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: E1123 20:20:34.472960 4726 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 23 20:20:34 crc kubenswrapper[4726]: E1123 20:20:34.473005 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3130c09b-7002-4448-bdd4-87b73f70ff99-cert podName:3130c09b-7002-4448-bdd4-87b73f70ff99 nodeName:}" failed. No retries permitted until 2025-11-23 20:20:34.972989759 +0000 UTC m=+743.122030715 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3130c09b-7002-4448-bdd4-87b73f70ff99-cert") pod "frr-k8s-webhook-server-6998585d5-mlnqp" (UID: "3130c09b-7002-4448-bdd4-87b73f70ff99") : secret "frr-k8s-webhook-server-cert" not found Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.473177 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-reloader\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.473339 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-conf\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: E1123 20:20:34.473399 4726 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 23 20:20:34 crc kubenswrapper[4726]: E1123 20:20:34.473427 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics-certs podName:87083cfa-76dd-42ca-a8e2-a3900f4f6bd5 nodeName:}" failed. No retries permitted until 2025-11-23 20:20:34.973419362 +0000 UTC m=+743.122460318 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics-certs") pod "frr-k8s-b8hdl" (UID: "87083cfa-76dd-42ca-a8e2-a3900f4f6bd5") : secret "frr-k8s-certs-secret" not found Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.473588 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.474549 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-frr-startup\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.497247 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xrmw\" (UniqueName: \"kubernetes.io/projected/3130c09b-7002-4448-bdd4-87b73f70ff99-kube-api-access-9xrmw\") pod \"frr-k8s-webhook-server-6998585d5-mlnqp\" (UID: \"3130c09b-7002-4448-bdd4-87b73f70ff99\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.509595 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pz5w\" (UniqueName: \"kubernetes.io/projected/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-kube-api-access-2pz5w\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.573035 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d99ab20e-6022-4d7e-8ff2-571cd0961a08-metallb-excludel2\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.573113 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.573157 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l4s5\" (UniqueName: \"kubernetes.io/projected/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-kube-api-access-8l4s5\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.573198 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-cert\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.573213 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-metrics-certs\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.573228 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fnmb\" (UniqueName: \"kubernetes.io/projected/d99ab20e-6022-4d7e-8ff2-571cd0961a08-kube-api-access-9fnmb\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.573261 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-metrics-certs\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: E1123 20:20:34.574065 4726 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 23 20:20:34 crc kubenswrapper[4726]: E1123 20:20:34.574143 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist podName:d99ab20e-6022-4d7e-8ff2-571cd0961a08 nodeName:}" failed. No retries permitted until 2025-11-23 20:20:35.074122432 +0000 UTC m=+743.223163388 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist") pod "speaker-zkvn5" (UID: "d99ab20e-6022-4d7e-8ff2-571cd0961a08") : secret "metallb-memberlist" not found Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.574533 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d99ab20e-6022-4d7e-8ff2-571cd0961a08-metallb-excludel2\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.579522 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-metrics-certs\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.579531 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-metrics-certs\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.579815 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.588755 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l4s5\" (UniqueName: \"kubernetes.io/projected/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-kube-api-access-8l4s5\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.592343 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fnmb\" (UniqueName: \"kubernetes.io/projected/d99ab20e-6022-4d7e-8ff2-571cd0961a08-kube-api-access-9fnmb\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.595629 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0-cert\") pod \"controller-6c7b4b5f48-2vb27\" (UID: \"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0\") " pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.684156 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.981458 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics-certs\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.981957 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3130c09b-7002-4448-bdd4-87b73f70ff99-cert\") pod \"frr-k8s-webhook-server-6998585d5-mlnqp\" (UID: \"3130c09b-7002-4448-bdd4-87b73f70ff99\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.987022 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87083cfa-76dd-42ca-a8e2-a3900f4f6bd5-metrics-certs\") pod \"frr-k8s-b8hdl\" (UID: \"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5\") " pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:34 crc kubenswrapper[4726]: I1123 20:20:34.987072 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3130c09b-7002-4448-bdd4-87b73f70ff99-cert\") pod \"frr-k8s-webhook-server-6998585d5-mlnqp\" (UID: \"3130c09b-7002-4448-bdd4-87b73f70ff99\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.092068 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:35 crc kubenswrapper[4726]: E1123 20:20:35.092536 4726 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 23 20:20:35 crc kubenswrapper[4726]: E1123 20:20:35.092714 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist podName:d99ab20e-6022-4d7e-8ff2-571cd0961a08 nodeName:}" failed. No retries permitted until 2025-11-23 20:20:36.092636435 +0000 UTC m=+744.241677421 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist") pod "speaker-zkvn5" (UID: "d99ab20e-6022-4d7e-8ff2-571cd0961a08") : secret "metallb-memberlist" not found Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.110728 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-2vb27"] Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.150137 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.178900 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.447697 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp"] Nov 23 20:20:35 crc kubenswrapper[4726]: W1123 20:20:35.453258 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3130c09b_7002_4448_bdd4_87b73f70ff99.slice/crio-b97300a21b481675812d711ea1bfb94089e19050433330a6961df39477db7c29 WatchSource:0}: Error finding container b97300a21b481675812d711ea1bfb94089e19050433330a6961df39477db7c29: Status 404 returned error can't find the container with id b97300a21b481675812d711ea1bfb94089e19050433330a6961df39477db7c29 Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.693056 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-2vb27" event={"ID":"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0","Type":"ContainerStarted","Data":"4089c075b94e1a76ee526b991a1af8a5e50b978b76a351b358b22acfa97a5ccd"} Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.693326 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-2vb27" event={"ID":"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0","Type":"ContainerStarted","Data":"3f4d2bf33d77a7487e1967ea8e74ce24be60841362ac095ea613f1a3abdb0b7d"} Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.693338 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-2vb27" event={"ID":"dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0","Type":"ContainerStarted","Data":"ed7caacba8d44555e124bd3938876975c4952049f285d6ccac83ee3a757ba42c"} Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.694425 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.695126 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerStarted","Data":"5e4c563ee4ce87f32cfa419f753e2e09784b3a14ca2ca3de3a0a95f7dbf9dea7"} Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.696191 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" event={"ID":"3130c09b-7002-4448-bdd4-87b73f70ff99","Type":"ContainerStarted","Data":"b97300a21b481675812d711ea1bfb94089e19050433330a6961df39477db7c29"} Nov 23 20:20:35 crc kubenswrapper[4726]: I1123 20:20:35.712935 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-2vb27" podStartSLOduration=1.7129127469999998 podStartE2EDuration="1.712912747s" podCreationTimestamp="2025-11-23 20:20:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:20:35.709741357 +0000 UTC m=+743.858782373" watchObservedRunningTime="2025-11-23 20:20:35.712912747 +0000 UTC m=+743.861953703" Nov 23 20:20:36 crc kubenswrapper[4726]: I1123 20:20:36.105439 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:36 crc kubenswrapper[4726]: I1123 20:20:36.114416 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d99ab20e-6022-4d7e-8ff2-571cd0961a08-memberlist\") pod \"speaker-zkvn5\" (UID: \"d99ab20e-6022-4d7e-8ff2-571cd0961a08\") " pod="metallb-system/speaker-zkvn5" Nov 23 20:20:36 crc kubenswrapper[4726]: I1123 20:20:36.173847 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zkvn5" Nov 23 20:20:36 crc kubenswrapper[4726]: W1123 20:20:36.198486 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd99ab20e_6022_4d7e_8ff2_571cd0961a08.slice/crio-8dadb2ba72a9c9585c9b9cd61c2b364cad5efb06aba80c1c3fdf7151d3b0b8e5 WatchSource:0}: Error finding container 8dadb2ba72a9c9585c9b9cd61c2b364cad5efb06aba80c1c3fdf7151d3b0b8e5: Status 404 returned error can't find the container with id 8dadb2ba72a9c9585c9b9cd61c2b364cad5efb06aba80c1c3fdf7151d3b0b8e5 Nov 23 20:20:36 crc kubenswrapper[4726]: I1123 20:20:36.706398 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zkvn5" event={"ID":"d99ab20e-6022-4d7e-8ff2-571cd0961a08","Type":"ContainerStarted","Data":"d672d433bde2753101097f22c3d22db986b8484d975ce4742fe9ab72686e712c"} Nov 23 20:20:36 crc kubenswrapper[4726]: I1123 20:20:36.706706 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zkvn5" event={"ID":"d99ab20e-6022-4d7e-8ff2-571cd0961a08","Type":"ContainerStarted","Data":"ba19728926ed65a8b01dd1b72be28cea29b2cbf276b4d824d553205f1aa05e33"} Nov 23 20:20:36 crc kubenswrapper[4726]: I1123 20:20:36.706716 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zkvn5" event={"ID":"d99ab20e-6022-4d7e-8ff2-571cd0961a08","Type":"ContainerStarted","Data":"8dadb2ba72a9c9585c9b9cd61c2b364cad5efb06aba80c1c3fdf7151d3b0b8e5"} Nov 23 20:20:36 crc kubenswrapper[4726]: I1123 20:20:36.706943 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-zkvn5" Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.307669 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-zkvn5" podStartSLOduration=3.307652946 podStartE2EDuration="3.307652946s" podCreationTimestamp="2025-11-23 20:20:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:20:36.725328838 +0000 UTC m=+744.874369814" watchObservedRunningTime="2025-11-23 20:20:37.307652946 +0000 UTC m=+745.456693902" Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.307831 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7ngz8"] Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.308094 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" podUID="b896681a-7b17-4f20-95a2-f0232ef91f3d" containerName="controller-manager" containerID="cri-o://446d9a54daf79eb2bbfe87a732ee034e94fbc9ba242853b355bca06640fc9111" gracePeriod=30 Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.408376 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8"] Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.408803 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" podUID="81862660-59a0-461e-b7b1-bf53c6e7a5e3" containerName="route-controller-manager" containerID="cri-o://8b55c8593ad2d04c18945d6c7ab850c97c49918e406bafde1f235f0d670dba85" gracePeriod=30 Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.719043 4726 generic.go:334] "Generic (PLEG): container finished" podID="b896681a-7b17-4f20-95a2-f0232ef91f3d" containerID="446d9a54daf79eb2bbfe87a732ee034e94fbc9ba242853b355bca06640fc9111" exitCode=0 Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.719177 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" event={"ID":"b896681a-7b17-4f20-95a2-f0232ef91f3d","Type":"ContainerDied","Data":"446d9a54daf79eb2bbfe87a732ee034e94fbc9ba242853b355bca06640fc9111"} Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.742246 4726 generic.go:334] "Generic (PLEG): container finished" podID="81862660-59a0-461e-b7b1-bf53c6e7a5e3" containerID="8b55c8593ad2d04c18945d6c7ab850c97c49918e406bafde1f235f0d670dba85" exitCode=0 Nov 23 20:20:37 crc kubenswrapper[4726]: I1123 20:20:37.742920 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" event={"ID":"81862660-59a0-461e-b7b1-bf53c6e7a5e3","Type":"ContainerDied","Data":"8b55c8593ad2d04c18945d6c7ab850c97c49918e406bafde1f235f0d670dba85"} Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.073486 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.235267 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.252061 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-client-ca\") pod \"b896681a-7b17-4f20-95a2-f0232ef91f3d\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.252153 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjc9n\" (UniqueName: \"kubernetes.io/projected/b896681a-7b17-4f20-95a2-f0232ef91f3d-kube-api-access-qjc9n\") pod \"b896681a-7b17-4f20-95a2-f0232ef91f3d\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.252189 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-config\") pod \"b896681a-7b17-4f20-95a2-f0232ef91f3d\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.252220 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-proxy-ca-bundles\") pod \"b896681a-7b17-4f20-95a2-f0232ef91f3d\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.252246 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b896681a-7b17-4f20-95a2-f0232ef91f3d-serving-cert\") pod \"b896681a-7b17-4f20-95a2-f0232ef91f3d\" (UID: \"b896681a-7b17-4f20-95a2-f0232ef91f3d\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.258600 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-config" (OuterVolumeSpecName: "config") pod "b896681a-7b17-4f20-95a2-f0232ef91f3d" (UID: "b896681a-7b17-4f20-95a2-f0232ef91f3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.259108 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b896681a-7b17-4f20-95a2-f0232ef91f3d-kube-api-access-qjc9n" (OuterVolumeSpecName: "kube-api-access-qjc9n") pod "b896681a-7b17-4f20-95a2-f0232ef91f3d" (UID: "b896681a-7b17-4f20-95a2-f0232ef91f3d"). InnerVolumeSpecName "kube-api-access-qjc9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.259116 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b896681a-7b17-4f20-95a2-f0232ef91f3d" (UID: "b896681a-7b17-4f20-95a2-f0232ef91f3d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.259430 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-client-ca" (OuterVolumeSpecName: "client-ca") pod "b896681a-7b17-4f20-95a2-f0232ef91f3d" (UID: "b896681a-7b17-4f20-95a2-f0232ef91f3d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.261129 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b896681a-7b17-4f20-95a2-f0232ef91f3d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b896681a-7b17-4f20-95a2-f0232ef91f3d" (UID: "b896681a-7b17-4f20-95a2-f0232ef91f3d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353050 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-config\") pod \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353170 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81862660-59a0-461e-b7b1-bf53c6e7a5e3-serving-cert\") pod \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353194 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-client-ca\") pod \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353215 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9bkd\" (UniqueName: \"kubernetes.io/projected/81862660-59a0-461e-b7b1-bf53c6e7a5e3-kube-api-access-c9bkd\") pod \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\" (UID: \"81862660-59a0-461e-b7b1-bf53c6e7a5e3\") " Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353413 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjc9n\" (UniqueName: \"kubernetes.io/projected/b896681a-7b17-4f20-95a2-f0232ef91f3d-kube-api-access-qjc9n\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353429 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353439 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353449 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b896681a-7b17-4f20-95a2-f0232ef91f3d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.353457 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b896681a-7b17-4f20-95a2-f0232ef91f3d-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.356235 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-client-ca" (OuterVolumeSpecName: "client-ca") pod "81862660-59a0-461e-b7b1-bf53c6e7a5e3" (UID: "81862660-59a0-461e-b7b1-bf53c6e7a5e3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.356310 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81862660-59a0-461e-b7b1-bf53c6e7a5e3-kube-api-access-c9bkd" (OuterVolumeSpecName: "kube-api-access-c9bkd") pod "81862660-59a0-461e-b7b1-bf53c6e7a5e3" (UID: "81862660-59a0-461e-b7b1-bf53c6e7a5e3"). InnerVolumeSpecName "kube-api-access-c9bkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.356635 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-config" (OuterVolumeSpecName: "config") pod "81862660-59a0-461e-b7b1-bf53c6e7a5e3" (UID: "81862660-59a0-461e-b7b1-bf53c6e7a5e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.365479 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81862660-59a0-461e-b7b1-bf53c6e7a5e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "81862660-59a0-461e-b7b1-bf53c6e7a5e3" (UID: "81862660-59a0-461e-b7b1-bf53c6e7a5e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.454439 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.454483 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81862660-59a0-461e-b7b1-bf53c6e7a5e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.454496 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/81862660-59a0-461e-b7b1-bf53c6e7a5e3-client-ca\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.454505 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9bkd\" (UniqueName: \"kubernetes.io/projected/81862660-59a0-461e-b7b1-bf53c6e7a5e3-kube-api-access-c9bkd\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.752330 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" event={"ID":"b896681a-7b17-4f20-95a2-f0232ef91f3d","Type":"ContainerDied","Data":"e9e1b21e538879a096bd3704701f09c0cb38dd8820f77b9e71af49d413c0e6d6"} Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.752390 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7ngz8" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.752402 4726 scope.go:117] "RemoveContainer" containerID="446d9a54daf79eb2bbfe87a732ee034e94fbc9ba242853b355bca06640fc9111" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.758072 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" event={"ID":"81862660-59a0-461e-b7b1-bf53c6e7a5e3","Type":"ContainerDied","Data":"040440aaa2e63c884c09d474b4fc95644b236c46f494608334925aa2f49539e1"} Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.758131 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.770541 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7ngz8"] Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.773244 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7ngz8"] Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.782257 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8"] Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.782638 4726 scope.go:117] "RemoveContainer" containerID="8b55c8593ad2d04c18945d6c7ab850c97c49918e406bafde1f235f0d670dba85" Nov 23 20:20:38 crc kubenswrapper[4726]: I1123 20:20:38.812665 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dnsz8"] Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.043086 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.043142 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.043217 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.046649 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ef166c526ceeb4bedeeddb339c120f70a8a458484a2d19fd740d49251ee0d41"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.046712 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://8ef166c526ceeb4bedeeddb339c120f70a8a458484a2d19fd740d49251ee0d41" gracePeriod=600 Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.381418 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp"] Nov 23 20:20:39 crc kubenswrapper[4726]: E1123 20:20:39.381931 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b896681a-7b17-4f20-95a2-f0232ef91f3d" containerName="controller-manager" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.381943 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b896681a-7b17-4f20-95a2-f0232ef91f3d" containerName="controller-manager" Nov 23 20:20:39 crc kubenswrapper[4726]: E1123 20:20:39.381960 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81862660-59a0-461e-b7b1-bf53c6e7a5e3" containerName="route-controller-manager" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.381967 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="81862660-59a0-461e-b7b1-bf53c6e7a5e3" containerName="route-controller-manager" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.382088 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="81862660-59a0-461e-b7b1-bf53c6e7a5e3" containerName="route-controller-manager" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.382103 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b896681a-7b17-4f20-95a2-f0232ef91f3d" containerName="controller-manager" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.382536 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.384536 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7fbb9896d7-59zr8"] Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.384776 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.385247 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.386692 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.386804 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.386883 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.387048 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.391539 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.398253 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.398463 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.398678 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.398781 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.398799 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.398950 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.406670 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp"] Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.412901 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7fbb9896d7-59zr8"] Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.431449 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.571893 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf9gl\" (UniqueName: \"kubernetes.io/projected/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-kube-api-access-jf9gl\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.571956 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-config\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.571985 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-serving-cert\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.572026 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-client-ca\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.572051 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtr54\" (UniqueName: \"kubernetes.io/projected/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-kube-api-access-dtr54\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.572072 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-serving-cert\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.572092 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-config\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.572106 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-proxy-ca-bundles\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.572132 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-client-ca\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-serving-cert\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672761 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-config\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672779 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-proxy-ca-bundles\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672807 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-client-ca\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672830 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf9gl\" (UniqueName: \"kubernetes.io/projected/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-kube-api-access-jf9gl\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-config\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672906 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-serving-cert\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672938 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-client-ca\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.672955 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtr54\" (UniqueName: \"kubernetes.io/projected/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-kube-api-access-dtr54\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.674141 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-client-ca\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.674167 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-client-ca\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.674422 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-config\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.675088 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-config\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.675156 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-proxy-ca-bundles\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.695586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-serving-cert\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.695624 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-serving-cert\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.700026 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtr54\" (UniqueName: \"kubernetes.io/projected/82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a-kube-api-access-dtr54\") pod \"controller-manager-7fbb9896d7-59zr8\" (UID: \"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a\") " pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.715323 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf9gl\" (UniqueName: \"kubernetes.io/projected/ba8e8cee-9a66-40b7-9c18-8e1b6eac021c-kube-api-access-jf9gl\") pod \"route-controller-manager-65789db6d5-9qbhp\" (UID: \"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c\") " pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.717174 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.783551 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="8ef166c526ceeb4bedeeddb339c120f70a8a458484a2d19fd740d49251ee0d41" exitCode=0 Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.783614 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"8ef166c526ceeb4bedeeddb339c120f70a8a458484a2d19fd740d49251ee0d41"} Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.783639 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"e0e5405a0e8a49ddd9351d6065e06af6eb07dcd086934b09caf4e017ae67cf3f"} Nov 23 20:20:39 crc kubenswrapper[4726]: I1123 20:20:39.783656 4726 scope.go:117] "RemoveContainer" containerID="5b42f4b47c86a8ff3e157d6e822c8e9bbf78256d76d3c430faf21561b98b7270" Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.000444 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.191436 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7fbb9896d7-59zr8"] Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.441974 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp"] Nov 23 20:20:40 crc kubenswrapper[4726]: W1123 20:20:40.462437 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba8e8cee_9a66_40b7_9c18_8e1b6eac021c.slice/crio-a8d6ace5d0c937170669327ce696b80fa9f93b1baa43784592bb9ed504a8e861 WatchSource:0}: Error finding container a8d6ace5d0c937170669327ce696b80fa9f93b1baa43784592bb9ed504a8e861: Status 404 returned error can't find the container with id a8d6ace5d0c937170669327ce696b80fa9f93b1baa43784592bb9ed504a8e861 Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.595534 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81862660-59a0-461e-b7b1-bf53c6e7a5e3" path="/var/lib/kubelet/pods/81862660-59a0-461e-b7b1-bf53c6e7a5e3/volumes" Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.596233 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b896681a-7b17-4f20-95a2-f0232ef91f3d" path="/var/lib/kubelet/pods/b896681a-7b17-4f20-95a2-f0232ef91f3d/volumes" Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.803845 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" event={"ID":"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c","Type":"ContainerStarted","Data":"78d6f93c83132f56f5a214ad212d46df70a709479e37dac0c76c833734154830"} Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.804229 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.804244 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" event={"ID":"ba8e8cee-9a66-40b7-9c18-8e1b6eac021c","Type":"ContainerStarted","Data":"a8d6ace5d0c937170669327ce696b80fa9f93b1baa43784592bb9ed504a8e861"} Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.807512 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" event={"ID":"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a","Type":"ContainerStarted","Data":"b111e40c98d1035bfbc62b2926afdcb87e48948e4b0941477a1c36fd1415d071"} Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.807575 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" event={"ID":"82dcf536-fa4d-4eb9-b6c3-296eabb4fa0a","Type":"ContainerStarted","Data":"08f296b675d1c646420d6ca6135256596424df3448480c8f1e5045e14ab0e39b"} Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.807765 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.816283 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.821895 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" podStartSLOduration=2.821859411 podStartE2EDuration="2.821859411s" podCreationTimestamp="2025-11-23 20:20:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:20:40.818367041 +0000 UTC m=+748.967408017" watchObservedRunningTime="2025-11-23 20:20:40.821859411 +0000 UTC m=+748.970900367" Nov 23 20:20:40 crc kubenswrapper[4726]: I1123 20:20:40.840643 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7fbb9896d7-59zr8" podStartSLOduration=3.840627858 podStartE2EDuration="3.840627858s" podCreationTimestamp="2025-11-23 20:20:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:20:40.836567212 +0000 UTC m=+748.985608168" watchObservedRunningTime="2025-11-23 20:20:40.840627858 +0000 UTC m=+748.989668814" Nov 23 20:20:41 crc kubenswrapper[4726]: I1123 20:20:41.125998 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-65789db6d5-9qbhp" Nov 23 20:20:45 crc kubenswrapper[4726]: I1123 20:20:45.848642 4726 generic.go:334] "Generic (PLEG): container finished" podID="87083cfa-76dd-42ca-a8e2-a3900f4f6bd5" containerID="b3ae5d188da467e35362902285f003aa49ea780e78d61ba69a773b33a8642c5f" exitCode=0 Nov 23 20:20:45 crc kubenswrapper[4726]: I1123 20:20:45.848752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerDied","Data":"b3ae5d188da467e35362902285f003aa49ea780e78d61ba69a773b33a8642c5f"} Nov 23 20:20:45 crc kubenswrapper[4726]: I1123 20:20:45.852616 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" event={"ID":"3130c09b-7002-4448-bdd4-87b73f70ff99","Type":"ContainerStarted","Data":"f0174f1d8e1639cef167a105a89ec405809e8200f2698c737127ed5ddaf176d6"} Nov 23 20:20:45 crc kubenswrapper[4726]: I1123 20:20:45.852779 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:45 crc kubenswrapper[4726]: I1123 20:20:45.909372 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" podStartSLOduration=2.197531341 podStartE2EDuration="11.909342631s" podCreationTimestamp="2025-11-23 20:20:34 +0000 UTC" firstStartedPulling="2025-11-23 20:20:35.456207835 +0000 UTC m=+743.605248791" lastFinishedPulling="2025-11-23 20:20:45.168019125 +0000 UTC m=+753.317060081" observedRunningTime="2025-11-23 20:20:45.904234255 +0000 UTC m=+754.053275241" watchObservedRunningTime="2025-11-23 20:20:45.909342631 +0000 UTC m=+754.058383587" Nov 23 20:20:46 crc kubenswrapper[4726]: I1123 20:20:46.181283 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-zkvn5" Nov 23 20:20:46 crc kubenswrapper[4726]: I1123 20:20:46.861143 4726 generic.go:334] "Generic (PLEG): container finished" podID="87083cfa-76dd-42ca-a8e2-a3900f4f6bd5" containerID="78eac23665fc62650114911eb3800742aa35cac646ef45d4cbccba91b3410920" exitCode=0 Nov 23 20:20:46 crc kubenswrapper[4726]: I1123 20:20:46.861238 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerDied","Data":"78eac23665fc62650114911eb3800742aa35cac646ef45d4cbccba91b3410920"} Nov 23 20:20:47 crc kubenswrapper[4726]: I1123 20:20:47.869951 4726 generic.go:334] "Generic (PLEG): container finished" podID="87083cfa-76dd-42ca-a8e2-a3900f4f6bd5" containerID="45fbcd4ba0caca38ae2492916f6807b1df17488da27bccb83974cbeb7166babf" exitCode=0 Nov 23 20:20:47 crc kubenswrapper[4726]: I1123 20:20:47.869990 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerDied","Data":"45fbcd4ba0caca38ae2492916f6807b1df17488da27bccb83974cbeb7166babf"} Nov 23 20:20:48 crc kubenswrapper[4726]: I1123 20:20:48.894294 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerStarted","Data":"c50ff58052ecf1463751d8c164abbe4eccbee44347b2ad3edf611e8ff9c52667"} Nov 23 20:20:48 crc kubenswrapper[4726]: I1123 20:20:48.894718 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerStarted","Data":"87302718d4966bcc5561e0ae4c2a5fb588ce5ca1358175e785a3988ec60a88d0"} Nov 23 20:20:48 crc kubenswrapper[4726]: I1123 20:20:48.894728 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerStarted","Data":"f03a4fe824121f222a3d578b9e96af8c7b889837b18be3b8fc28f775556f2736"} Nov 23 20:20:48 crc kubenswrapper[4726]: I1123 20:20:48.894737 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerStarted","Data":"20e0127cf67ff5cf98541728ed35a77dda3c44a7484f74f0f8dd6ec3f25b12a5"} Nov 23 20:20:48 crc kubenswrapper[4726]: I1123 20:20:48.894746 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerStarted","Data":"25acf0a0be881022386ba162487873b515d2d9515609371b87ee2456d6906725"} Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.248527 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-52k2c"] Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.249389 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-52k2c" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.272335 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.272491 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-52k2c"] Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.272991 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.409978 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t5qp\" (UniqueName: \"kubernetes.io/projected/af4e8328-2292-460c-bcec-801528e6e778-kube-api-access-5t5qp\") pod \"openstack-operator-index-52k2c\" (UID: \"af4e8328-2292-460c-bcec-801528e6e778\") " pod="openstack-operators/openstack-operator-index-52k2c" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.511537 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t5qp\" (UniqueName: \"kubernetes.io/projected/af4e8328-2292-460c-bcec-801528e6e778-kube-api-access-5t5qp\") pod \"openstack-operator-index-52k2c\" (UID: \"af4e8328-2292-460c-bcec-801528e6e778\") " pod="openstack-operators/openstack-operator-index-52k2c" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.534571 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t5qp\" (UniqueName: \"kubernetes.io/projected/af4e8328-2292-460c-bcec-801528e6e778-kube-api-access-5t5qp\") pod \"openstack-operator-index-52k2c\" (UID: \"af4e8328-2292-460c-bcec-801528e6e778\") " pod="openstack-operators/openstack-operator-index-52k2c" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.581811 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-52k2c" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.602174 4726 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.905106 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-b8hdl" event={"ID":"87083cfa-76dd-42ca-a8e2-a3900f4f6bd5","Type":"ContainerStarted","Data":"08267b47832a09ffe1c43bf26e8937bdaa8e9846e21a51bac0e0ee81a2d8a072"} Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.905472 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:49 crc kubenswrapper[4726]: I1123 20:20:49.925142 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-b8hdl" podStartSLOduration=6.12193329 podStartE2EDuration="15.925126934s" podCreationTimestamp="2025-11-23 20:20:34 +0000 UTC" firstStartedPulling="2025-11-23 20:20:35.384781442 +0000 UTC m=+743.533822398" lastFinishedPulling="2025-11-23 20:20:45.187975076 +0000 UTC m=+753.337016042" observedRunningTime="2025-11-23 20:20:49.921356426 +0000 UTC m=+758.070397392" watchObservedRunningTime="2025-11-23 20:20:49.925126934 +0000 UTC m=+758.074167890" Nov 23 20:20:50 crc kubenswrapper[4726]: I1123 20:20:50.155913 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-52k2c"] Nov 23 20:20:50 crc kubenswrapper[4726]: I1123 20:20:50.180057 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:50 crc kubenswrapper[4726]: I1123 20:20:50.220494 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:20:50 crc kubenswrapper[4726]: I1123 20:20:50.910061 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-52k2c" event={"ID":"af4e8328-2292-460c-bcec-801528e6e778","Type":"ContainerStarted","Data":"7b09d42f5a81141b8914bd738d4e3ae2a4cc918b0fe58c56a4de03b3b924a79a"} Nov 23 20:20:51 crc kubenswrapper[4726]: I1123 20:20:51.916996 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-52k2c" event={"ID":"af4e8328-2292-460c-bcec-801528e6e778","Type":"ContainerStarted","Data":"17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154"} Nov 23 20:20:51 crc kubenswrapper[4726]: I1123 20:20:51.933767 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-52k2c" podStartSLOduration=2.229006152 podStartE2EDuration="2.933746252s" podCreationTimestamp="2025-11-23 20:20:49 +0000 UTC" firstStartedPulling="2025-11-23 20:20:50.15222624 +0000 UTC m=+758.301267196" lastFinishedPulling="2025-11-23 20:20:50.85696634 +0000 UTC m=+759.006007296" observedRunningTime="2025-11-23 20:20:51.931049434 +0000 UTC m=+760.080090430" watchObservedRunningTime="2025-11-23 20:20:51.933746252 +0000 UTC m=+760.082787208" Nov 23 20:20:52 crc kubenswrapper[4726]: I1123 20:20:52.625985 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-52k2c"] Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.231893 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-whxfc"] Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.233207 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.239435 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-n2fz6" Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.242750 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-whxfc"] Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.369832 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxbgh\" (UniqueName: \"kubernetes.io/projected/224802b7-2607-4e77-95e0-d7e02023c20b-kube-api-access-sxbgh\") pod \"openstack-operator-index-whxfc\" (UID: \"224802b7-2607-4e77-95e0-d7e02023c20b\") " pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.471712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxbgh\" (UniqueName: \"kubernetes.io/projected/224802b7-2607-4e77-95e0-d7e02023c20b-kube-api-access-sxbgh\") pod \"openstack-operator-index-whxfc\" (UID: \"224802b7-2607-4e77-95e0-d7e02023c20b\") " pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.495991 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxbgh\" (UniqueName: \"kubernetes.io/projected/224802b7-2607-4e77-95e0-d7e02023c20b-kube-api-access-sxbgh\") pod \"openstack-operator-index-whxfc\" (UID: \"224802b7-2607-4e77-95e0-d7e02023c20b\") " pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.577708 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:20:53 crc kubenswrapper[4726]: I1123 20:20:53.926585 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-52k2c" podUID="af4e8328-2292-460c-bcec-801528e6e778" containerName="registry-server" containerID="cri-o://17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154" gracePeriod=2 Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.021381 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-whxfc"] Nov 23 20:20:54 crc kubenswrapper[4726]: W1123 20:20:54.076971 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod224802b7_2607_4e77_95e0_d7e02023c20b.slice/crio-7da987638a3edff47770c7291a90950ce22acf78613b30fb5b99b28fd10489fa WatchSource:0}: Error finding container 7da987638a3edff47770c7291a90950ce22acf78613b30fb5b99b28fd10489fa: Status 404 returned error can't find the container with id 7da987638a3edff47770c7291a90950ce22acf78613b30fb5b99b28fd10489fa Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.355048 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-52k2c" Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.487345 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t5qp\" (UniqueName: \"kubernetes.io/projected/af4e8328-2292-460c-bcec-801528e6e778-kube-api-access-5t5qp\") pod \"af4e8328-2292-460c-bcec-801528e6e778\" (UID: \"af4e8328-2292-460c-bcec-801528e6e778\") " Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.496699 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af4e8328-2292-460c-bcec-801528e6e778-kube-api-access-5t5qp" (OuterVolumeSpecName: "kube-api-access-5t5qp") pod "af4e8328-2292-460c-bcec-801528e6e778" (UID: "af4e8328-2292-460c-bcec-801528e6e778"). InnerVolumeSpecName "kube-api-access-5t5qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.590811 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t5qp\" (UniqueName: \"kubernetes.io/projected/af4e8328-2292-460c-bcec-801528e6e778-kube-api-access-5t5qp\") on node \"crc\" DevicePath \"\"" Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.688057 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-2vb27" Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.935838 4726 generic.go:334] "Generic (PLEG): container finished" podID="af4e8328-2292-460c-bcec-801528e6e778" containerID="17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154" exitCode=0 Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.935997 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-52k2c" Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.936054 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-52k2c" event={"ID":"af4e8328-2292-460c-bcec-801528e6e778","Type":"ContainerDied","Data":"17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154"} Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.936109 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-52k2c" event={"ID":"af4e8328-2292-460c-bcec-801528e6e778","Type":"ContainerDied","Data":"7b09d42f5a81141b8914bd738d4e3ae2a4cc918b0fe58c56a4de03b3b924a79a"} Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.936130 4726 scope.go:117] "RemoveContainer" containerID="17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154" Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.942230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-whxfc" event={"ID":"224802b7-2607-4e77-95e0-d7e02023c20b","Type":"ContainerStarted","Data":"8bc5a0d2698099fd6dcdee46c237928b88bdc708e07b770d4041fdea06c9526a"} Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.942961 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-whxfc" event={"ID":"224802b7-2607-4e77-95e0-d7e02023c20b","Type":"ContainerStarted","Data":"7da987638a3edff47770c7291a90950ce22acf78613b30fb5b99b28fd10489fa"} Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.962191 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-52k2c"] Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.969252 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-52k2c"] Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.971033 4726 scope.go:117] "RemoveContainer" containerID="17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154" Nov 23 20:20:54 crc kubenswrapper[4726]: E1123 20:20:54.971771 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154\": container with ID starting with 17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154 not found: ID does not exist" containerID="17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154" Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.971807 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154"} err="failed to get container status \"17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154\": rpc error: code = NotFound desc = could not find container \"17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154\": container with ID starting with 17c07b3b7dd919fd028576e2c0d2346307bf9c13500344af702052f5b855c154 not found: ID does not exist" Nov 23 20:20:54 crc kubenswrapper[4726]: I1123 20:20:54.981043 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-whxfc" podStartSLOduration=1.53460284 podStartE2EDuration="1.98101535s" podCreationTimestamp="2025-11-23 20:20:53 +0000 UTC" firstStartedPulling="2025-11-23 20:20:54.080088698 +0000 UTC m=+762.229129664" lastFinishedPulling="2025-11-23 20:20:54.526501198 +0000 UTC m=+762.675542174" observedRunningTime="2025-11-23 20:20:54.977375276 +0000 UTC m=+763.126416242" watchObservedRunningTime="2025-11-23 20:20:54.98101535 +0000 UTC m=+763.130056346" Nov 23 20:20:55 crc kubenswrapper[4726]: I1123 20:20:55.155115 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-mlnqp" Nov 23 20:20:56 crc kubenswrapper[4726]: I1123 20:20:56.631765 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af4e8328-2292-460c-bcec-801528e6e778" path="/var/lib/kubelet/pods/af4e8328-2292-460c-bcec-801528e6e778/volumes" Nov 23 20:21:03 crc kubenswrapper[4726]: I1123 20:21:03.578290 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:21:03 crc kubenswrapper[4726]: I1123 20:21:03.579134 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:21:03 crc kubenswrapper[4726]: I1123 20:21:03.626187 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.066461 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-whxfc" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.686447 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr"] Nov 23 20:21:04 crc kubenswrapper[4726]: E1123 20:21:04.687160 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af4e8328-2292-460c-bcec-801528e6e778" containerName="registry-server" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.687182 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="af4e8328-2292-460c-bcec-801528e6e778" containerName="registry-server" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.687402 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="af4e8328-2292-460c-bcec-801528e6e778" containerName="registry-server" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.688709 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.693610 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-l7jgr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.702440 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr"] Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.851205 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.851281 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp2rf\" (UniqueName: \"kubernetes.io/projected/6a77b1d9-7f6f-4ea1-847a-393ba021319f-kube-api-access-vp2rf\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.851478 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.953297 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.953395 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp2rf\" (UniqueName: \"kubernetes.io/projected/6a77b1d9-7f6f-4ea1-847a-393ba021319f-kube-api-access-vp2rf\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.953444 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.953726 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.954309 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:04 crc kubenswrapper[4726]: I1123 20:21:04.978727 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp2rf\" (UniqueName: \"kubernetes.io/projected/6a77b1d9-7f6f-4ea1-847a-393ba021319f-kube-api-access-vp2rf\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:05 crc kubenswrapper[4726]: I1123 20:21:05.030789 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:05 crc kubenswrapper[4726]: I1123 20:21:05.182074 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-b8hdl" Nov 23 20:21:05 crc kubenswrapper[4726]: I1123 20:21:05.521560 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr"] Nov 23 20:21:05 crc kubenswrapper[4726]: W1123 20:21:05.525167 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a77b1d9_7f6f_4ea1_847a_393ba021319f.slice/crio-bc144ebe2df5fcfdde44aea2009837bb5dbe1bf03c3c88ecf77a15b885cd3a3e WatchSource:0}: Error finding container bc144ebe2df5fcfdde44aea2009837bb5dbe1bf03c3c88ecf77a15b885cd3a3e: Status 404 returned error can't find the container with id bc144ebe2df5fcfdde44aea2009837bb5dbe1bf03c3c88ecf77a15b885cd3a3e Nov 23 20:21:06 crc kubenswrapper[4726]: I1123 20:21:06.040728 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerID="6b3d048e221e02ef9a43e04b4ba117e03e824acb5aae610da4ba3d817ce15022" exitCode=0 Nov 23 20:21:06 crc kubenswrapper[4726]: I1123 20:21:06.040800 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" event={"ID":"6a77b1d9-7f6f-4ea1-847a-393ba021319f","Type":"ContainerDied","Data":"6b3d048e221e02ef9a43e04b4ba117e03e824acb5aae610da4ba3d817ce15022"} Nov 23 20:21:06 crc kubenswrapper[4726]: I1123 20:21:06.041108 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" event={"ID":"6a77b1d9-7f6f-4ea1-847a-393ba021319f","Type":"ContainerStarted","Data":"bc144ebe2df5fcfdde44aea2009837bb5dbe1bf03c3c88ecf77a15b885cd3a3e"} Nov 23 20:21:07 crc kubenswrapper[4726]: I1123 20:21:07.048910 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerID="ab5569eed52a288f7cfe16c3a1be6f93919c6d53371434878befdabff346edee" exitCode=0 Nov 23 20:21:07 crc kubenswrapper[4726]: I1123 20:21:07.048956 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" event={"ID":"6a77b1d9-7f6f-4ea1-847a-393ba021319f","Type":"ContainerDied","Data":"ab5569eed52a288f7cfe16c3a1be6f93919c6d53371434878befdabff346edee"} Nov 23 20:21:08 crc kubenswrapper[4726]: I1123 20:21:08.060958 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerID="b5c3e5cb2338b22fc686e3a71ea6dd05cbbf5c207d84e188b9bf84564f2c07e7" exitCode=0 Nov 23 20:21:08 crc kubenswrapper[4726]: I1123 20:21:08.061027 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" event={"ID":"6a77b1d9-7f6f-4ea1-847a-393ba021319f","Type":"ContainerDied","Data":"b5c3e5cb2338b22fc686e3a71ea6dd05cbbf5c207d84e188b9bf84564f2c07e7"} Nov 23 20:21:08 crc kubenswrapper[4726]: I1123 20:21:08.843235 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4pqht"] Nov 23 20:21:08 crc kubenswrapper[4726]: I1123 20:21:08.847787 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:08 crc kubenswrapper[4726]: I1123 20:21:08.856992 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4pqht"] Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.015006 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-utilities\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.015083 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-catalog-content\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.015315 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxt9d\" (UniqueName: \"kubernetes.io/projected/15fc8248-951f-4041-a122-ff4f3e30050b-kube-api-access-cxt9d\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.117022 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-catalog-content\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.117073 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-utilities\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.117126 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxt9d\" (UniqueName: \"kubernetes.io/projected/15fc8248-951f-4041-a122-ff4f3e30050b-kube-api-access-cxt9d\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.117971 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-catalog-content\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.118123 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-utilities\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.140091 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxt9d\" (UniqueName: \"kubernetes.io/projected/15fc8248-951f-4041-a122-ff4f3e30050b-kube-api-access-cxt9d\") pod \"community-operators-4pqht\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.200697 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.445981 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.533510 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-util\") pod \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.533576 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-bundle\") pod \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.533723 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp2rf\" (UniqueName: \"kubernetes.io/projected/6a77b1d9-7f6f-4ea1-847a-393ba021319f-kube-api-access-vp2rf\") pod \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\" (UID: \"6a77b1d9-7f6f-4ea1-847a-393ba021319f\") " Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.534321 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-bundle" (OuterVolumeSpecName: "bundle") pod "6a77b1d9-7f6f-4ea1-847a-393ba021319f" (UID: "6a77b1d9-7f6f-4ea1-847a-393ba021319f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.540142 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a77b1d9-7f6f-4ea1-847a-393ba021319f-kube-api-access-vp2rf" (OuterVolumeSpecName: "kube-api-access-vp2rf") pod "6a77b1d9-7f6f-4ea1-847a-393ba021319f" (UID: "6a77b1d9-7f6f-4ea1-847a-393ba021319f"). InnerVolumeSpecName "kube-api-access-vp2rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.552006 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-util" (OuterVolumeSpecName: "util") pod "6a77b1d9-7f6f-4ea1-847a-393ba021319f" (UID: "6a77b1d9-7f6f-4ea1-847a-393ba021319f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.635074 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp2rf\" (UniqueName: \"kubernetes.io/projected/6a77b1d9-7f6f-4ea1-847a-393ba021319f-kube-api-access-vp2rf\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.635100 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-util\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.635111 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a77b1d9-7f6f-4ea1-847a-393ba021319f-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:09 crc kubenswrapper[4726]: I1123 20:21:09.670501 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4pqht"] Nov 23 20:21:10 crc kubenswrapper[4726]: I1123 20:21:10.078292 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" event={"ID":"6a77b1d9-7f6f-4ea1-847a-393ba021319f","Type":"ContainerDied","Data":"bc144ebe2df5fcfdde44aea2009837bb5dbe1bf03c3c88ecf77a15b885cd3a3e"} Nov 23 20:21:10 crc kubenswrapper[4726]: I1123 20:21:10.078679 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc144ebe2df5fcfdde44aea2009837bb5dbe1bf03c3c88ecf77a15b885cd3a3e" Nov 23 20:21:10 crc kubenswrapper[4726]: I1123 20:21:10.078343 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr" Nov 23 20:21:10 crc kubenswrapper[4726]: I1123 20:21:10.081084 4726 generic.go:334] "Generic (PLEG): container finished" podID="15fc8248-951f-4041-a122-ff4f3e30050b" containerID="20a43ef31eaf6ee3fe2af6a7c38f5172cb85af4cd44a152dcd81a05fd95f8453" exitCode=0 Nov 23 20:21:10 crc kubenswrapper[4726]: I1123 20:21:10.081131 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pqht" event={"ID":"15fc8248-951f-4041-a122-ff4f3e30050b","Type":"ContainerDied","Data":"20a43ef31eaf6ee3fe2af6a7c38f5172cb85af4cd44a152dcd81a05fd95f8453"} Nov 23 20:21:10 crc kubenswrapper[4726]: I1123 20:21:10.081160 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pqht" event={"ID":"15fc8248-951f-4041-a122-ff4f3e30050b","Type":"ContainerStarted","Data":"968e7227ac9e458bcc11a8dc7fea2ab40feee3e9de9b84fbf49f031d4c2d44fd"} Nov 23 20:21:11 crc kubenswrapper[4726]: I1123 20:21:11.089802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pqht" event={"ID":"15fc8248-951f-4041-a122-ff4f3e30050b","Type":"ContainerStarted","Data":"12dae09b369d39fc8283c67b6e3897ba96fb2c7b98ddbeb7a65c1d0160a048de"} Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.039669 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rq6nk"] Nov 23 20:21:12 crc kubenswrapper[4726]: E1123 20:21:12.040429 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerName="extract" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.040459 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerName="extract" Nov 23 20:21:12 crc kubenswrapper[4726]: E1123 20:21:12.040487 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerName="pull" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.040501 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerName="pull" Nov 23 20:21:12 crc kubenswrapper[4726]: E1123 20:21:12.040522 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerName="util" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.040534 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerName="util" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.040763 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a77b1d9-7f6f-4ea1-847a-393ba021319f" containerName="extract" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.042319 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.056984 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rq6nk"] Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.098351 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pqht" event={"ID":"15fc8248-951f-4041-a122-ff4f3e30050b","Type":"ContainerDied","Data":"12dae09b369d39fc8283c67b6e3897ba96fb2c7b98ddbeb7a65c1d0160a048de"} Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.098213 4726 generic.go:334] "Generic (PLEG): container finished" podID="15fc8248-951f-4041-a122-ff4f3e30050b" containerID="12dae09b369d39fc8283c67b6e3897ba96fb2c7b98ddbeb7a65c1d0160a048de" exitCode=0 Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.182470 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-catalog-content\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.182537 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmjln\" (UniqueName: \"kubernetes.io/projected/9338021a-a0ec-4b95-ad96-d4e95439b47b-kube-api-access-cmjln\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.182579 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-utilities\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.284035 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-catalog-content\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.284095 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmjln\" (UniqueName: \"kubernetes.io/projected/9338021a-a0ec-4b95-ad96-d4e95439b47b-kube-api-access-cmjln\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.284142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-utilities\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.284727 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-catalog-content\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.284738 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-utilities\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.313643 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmjln\" (UniqueName: \"kubernetes.io/projected/9338021a-a0ec-4b95-ad96-d4e95439b47b-kube-api-access-cmjln\") pod \"redhat-marketplace-rq6nk\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.362464 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:12 crc kubenswrapper[4726]: I1123 20:21:12.803897 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rq6nk"] Nov 23 20:21:13 crc kubenswrapper[4726]: I1123 20:21:13.109142 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pqht" event={"ID":"15fc8248-951f-4041-a122-ff4f3e30050b","Type":"ContainerStarted","Data":"925ec625f921f425b67467e2d98c6a00ee4b102e305d3acc944deccadb15c128"} Nov 23 20:21:13 crc kubenswrapper[4726]: I1123 20:21:13.111672 4726 generic.go:334] "Generic (PLEG): container finished" podID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerID="236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4" exitCode=0 Nov 23 20:21:13 crc kubenswrapper[4726]: I1123 20:21:13.111734 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq6nk" event={"ID":"9338021a-a0ec-4b95-ad96-d4e95439b47b","Type":"ContainerDied","Data":"236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4"} Nov 23 20:21:13 crc kubenswrapper[4726]: I1123 20:21:13.111785 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq6nk" event={"ID":"9338021a-a0ec-4b95-ad96-d4e95439b47b","Type":"ContainerStarted","Data":"bb536fa93c36ad2bbd8fe03c82c276b628eb6a42d417b0498211deb2ad577b92"} Nov 23 20:21:13 crc kubenswrapper[4726]: I1123 20:21:13.131042 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4pqht" podStartSLOduration=2.648225277 podStartE2EDuration="5.131012578s" podCreationTimestamp="2025-11-23 20:21:08 +0000 UTC" firstStartedPulling="2025-11-23 20:21:10.083824632 +0000 UTC m=+778.232865618" lastFinishedPulling="2025-11-23 20:21:12.566611963 +0000 UTC m=+780.715652919" observedRunningTime="2025-11-23 20:21:13.127553479 +0000 UTC m=+781.276594495" watchObservedRunningTime="2025-11-23 20:21:13.131012578 +0000 UTC m=+781.280053604" Nov 23 20:21:14 crc kubenswrapper[4726]: I1123 20:21:14.121051 4726 generic.go:334] "Generic (PLEG): container finished" podID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerID="d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7" exitCode=0 Nov 23 20:21:14 crc kubenswrapper[4726]: I1123 20:21:14.121097 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq6nk" event={"ID":"9338021a-a0ec-4b95-ad96-d4e95439b47b","Type":"ContainerDied","Data":"d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7"} Nov 23 20:21:15 crc kubenswrapper[4726]: I1123 20:21:15.128825 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq6nk" event={"ID":"9338021a-a0ec-4b95-ad96-d4e95439b47b","Type":"ContainerStarted","Data":"78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2"} Nov 23 20:21:15 crc kubenswrapper[4726]: I1123 20:21:15.160162 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rq6nk" podStartSLOduration=1.740726598 podStartE2EDuration="3.160145257s" podCreationTimestamp="2025-11-23 20:21:12 +0000 UTC" firstStartedPulling="2025-11-23 20:21:13.113784855 +0000 UTC m=+781.262825831" lastFinishedPulling="2025-11-23 20:21:14.533203534 +0000 UTC m=+782.682244490" observedRunningTime="2025-11-23 20:21:15.159587921 +0000 UTC m=+783.308628887" watchObservedRunningTime="2025-11-23 20:21:15.160145257 +0000 UTC m=+783.309186213" Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.196157 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf"] Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.197020 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.198953 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-zv5ch" Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.233306 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf"] Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.333554 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjpjj\" (UniqueName: \"kubernetes.io/projected/2a88699d-09b3-458d-bbbd-03834f3c6978-kube-api-access-sjpjj\") pod \"openstack-operator-controller-operator-8486c7f98b-v5bmf\" (UID: \"2a88699d-09b3-458d-bbbd-03834f3c6978\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.435129 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjpjj\" (UniqueName: \"kubernetes.io/projected/2a88699d-09b3-458d-bbbd-03834f3c6978-kube-api-access-sjpjj\") pod \"openstack-operator-controller-operator-8486c7f98b-v5bmf\" (UID: \"2a88699d-09b3-458d-bbbd-03834f3c6978\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.458269 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjpjj\" (UniqueName: \"kubernetes.io/projected/2a88699d-09b3-458d-bbbd-03834f3c6978-kube-api-access-sjpjj\") pod \"openstack-operator-controller-operator-8486c7f98b-v5bmf\" (UID: \"2a88699d-09b3-458d-bbbd-03834f3c6978\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.511842 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" Nov 23 20:21:16 crc kubenswrapper[4726]: I1123 20:21:16.944453 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf"] Nov 23 20:21:17 crc kubenswrapper[4726]: I1123 20:21:17.138077 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" event={"ID":"2a88699d-09b3-458d-bbbd-03834f3c6978","Type":"ContainerStarted","Data":"45185376cc6294b82e291e35946805669375eae1893cd60a037737e49932ee64"} Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.231418 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zx5vk"] Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.232636 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.237293 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zx5vk"] Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.359744 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-catalog-content\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.359799 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jkwm\" (UniqueName: \"kubernetes.io/projected/71bfcdcb-9312-4684-9b7c-3428aec8118a-kube-api-access-9jkwm\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.359845 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-utilities\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.466217 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-catalog-content\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.466255 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jkwm\" (UniqueName: \"kubernetes.io/projected/71bfcdcb-9312-4684-9b7c-3428aec8118a-kube-api-access-9jkwm\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.466273 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-utilities\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.466723 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-utilities\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.466907 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-catalog-content\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.484131 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jkwm\" (UniqueName: \"kubernetes.io/projected/71bfcdcb-9312-4684-9b7c-3428aec8118a-kube-api-access-9jkwm\") pod \"redhat-operators-zx5vk\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:18 crc kubenswrapper[4726]: I1123 20:21:18.565956 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:19 crc kubenswrapper[4726]: I1123 20:21:19.201608 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:19 crc kubenswrapper[4726]: I1123 20:21:19.201672 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:19 crc kubenswrapper[4726]: I1123 20:21:19.242370 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:20 crc kubenswrapper[4726]: I1123 20:21:20.192861 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:21 crc kubenswrapper[4726]: I1123 20:21:21.732066 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zx5vk"] Nov 23 20:21:21 crc kubenswrapper[4726]: W1123 20:21:21.740966 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bfcdcb_9312_4684_9b7c_3428aec8118a.slice/crio-12f743ceac5a658bc1aee269999f7d1706aa196b7fd8a5e592b4bc4d4ce69b43 WatchSource:0}: Error finding container 12f743ceac5a658bc1aee269999f7d1706aa196b7fd8a5e592b4bc4d4ce69b43: Status 404 returned error can't find the container with id 12f743ceac5a658bc1aee269999f7d1706aa196b7fd8a5e592b4bc4d4ce69b43 Nov 23 20:21:22 crc kubenswrapper[4726]: I1123 20:21:22.185691 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" event={"ID":"2a88699d-09b3-458d-bbbd-03834f3c6978","Type":"ContainerStarted","Data":"0dc0159cbc1a87cebfb751d1384e5d56d3d6f0f1c6c47ac6f36773344f8b378d"} Nov 23 20:21:22 crc kubenswrapper[4726]: I1123 20:21:22.188088 4726 generic.go:334] "Generic (PLEG): container finished" podID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerID="8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7" exitCode=0 Nov 23 20:21:22 crc kubenswrapper[4726]: I1123 20:21:22.188128 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx5vk" event={"ID":"71bfcdcb-9312-4684-9b7c-3428aec8118a","Type":"ContainerDied","Data":"8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7"} Nov 23 20:21:22 crc kubenswrapper[4726]: I1123 20:21:22.188173 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx5vk" event={"ID":"71bfcdcb-9312-4684-9b7c-3428aec8118a","Type":"ContainerStarted","Data":"12f743ceac5a658bc1aee269999f7d1706aa196b7fd8a5e592b4bc4d4ce69b43"} Nov 23 20:21:22 crc kubenswrapper[4726]: I1123 20:21:22.363233 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:22 crc kubenswrapper[4726]: I1123 20:21:22.363277 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:22 crc kubenswrapper[4726]: I1123 20:21:22.478502 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:23 crc kubenswrapper[4726]: I1123 20:21:23.264245 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:24 crc kubenswrapper[4726]: I1123 20:21:24.818834 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4pqht"] Nov 23 20:21:24 crc kubenswrapper[4726]: I1123 20:21:24.819123 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4pqht" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" containerName="registry-server" containerID="cri-o://925ec625f921f425b67467e2d98c6a00ee4b102e305d3acc944deccadb15c128" gracePeriod=2 Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.205912 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" event={"ID":"2a88699d-09b3-458d-bbbd-03834f3c6978","Type":"ContainerStarted","Data":"0c188dc4803ebd6ceb86620ce477e392c083487eb09558af2ff7d392a6c5104a"} Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.206616 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.208082 4726 generic.go:334] "Generic (PLEG): container finished" podID="15fc8248-951f-4041-a122-ff4f3e30050b" containerID="925ec625f921f425b67467e2d98c6a00ee4b102e305d3acc944deccadb15c128" exitCode=0 Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.208189 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pqht" event={"ID":"15fc8248-951f-4041-a122-ff4f3e30050b","Type":"ContainerDied","Data":"925ec625f921f425b67467e2d98c6a00ee4b102e305d3acc944deccadb15c128"} Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.208279 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4pqht" event={"ID":"15fc8248-951f-4041-a122-ff4f3e30050b","Type":"ContainerDied","Data":"968e7227ac9e458bcc11a8dc7fea2ab40feee3e9de9b84fbf49f031d4c2d44fd"} Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.208339 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="968e7227ac9e458bcc11a8dc7fea2ab40feee3e9de9b84fbf49f031d4c2d44fd" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.212710 4726 generic.go:334] "Generic (PLEG): container finished" podID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerID="209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac" exitCode=0 Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.212812 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx5vk" event={"ID":"71bfcdcb-9312-4684-9b7c-3428aec8118a","Type":"ContainerDied","Data":"209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac"} Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.240897 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" podStartSLOduration=2.100506526 podStartE2EDuration="9.240880091s" podCreationTimestamp="2025-11-23 20:21:16 +0000 UTC" firstStartedPulling="2025-11-23 20:21:16.950405244 +0000 UTC m=+785.099446200" lastFinishedPulling="2025-11-23 20:21:24.090778799 +0000 UTC m=+792.239819765" observedRunningTime="2025-11-23 20:21:25.238347978 +0000 UTC m=+793.387388964" watchObservedRunningTime="2025-11-23 20:21:25.240880091 +0000 UTC m=+793.389921047" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.262459 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.384425 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxt9d\" (UniqueName: \"kubernetes.io/projected/15fc8248-951f-4041-a122-ff4f3e30050b-kube-api-access-cxt9d\") pod \"15fc8248-951f-4041-a122-ff4f3e30050b\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.384485 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-catalog-content\") pod \"15fc8248-951f-4041-a122-ff4f3e30050b\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.384531 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-utilities\") pod \"15fc8248-951f-4041-a122-ff4f3e30050b\" (UID: \"15fc8248-951f-4041-a122-ff4f3e30050b\") " Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.385806 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-utilities" (OuterVolumeSpecName: "utilities") pod "15fc8248-951f-4041-a122-ff4f3e30050b" (UID: "15fc8248-951f-4041-a122-ff4f3e30050b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.389736 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15fc8248-951f-4041-a122-ff4f3e30050b-kube-api-access-cxt9d" (OuterVolumeSpecName: "kube-api-access-cxt9d") pod "15fc8248-951f-4041-a122-ff4f3e30050b" (UID: "15fc8248-951f-4041-a122-ff4f3e30050b"). InnerVolumeSpecName "kube-api-access-cxt9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.429804 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15fc8248-951f-4041-a122-ff4f3e30050b" (UID: "15fc8248-951f-4041-a122-ff4f3e30050b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.485890 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxt9d\" (UniqueName: \"kubernetes.io/projected/15fc8248-951f-4041-a122-ff4f3e30050b-kube-api-access-cxt9d\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.485928 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:25 crc kubenswrapper[4726]: I1123 20:21:25.485940 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fc8248-951f-4041-a122-ff4f3e30050b-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:26 crc kubenswrapper[4726]: I1123 20:21:26.221840 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx5vk" event={"ID":"71bfcdcb-9312-4684-9b7c-3428aec8118a","Type":"ContainerStarted","Data":"aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37"} Nov 23 20:21:26 crc kubenswrapper[4726]: I1123 20:21:26.222251 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4pqht" Nov 23 20:21:26 crc kubenswrapper[4726]: I1123 20:21:26.223658 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-v5bmf" Nov 23 20:21:26 crc kubenswrapper[4726]: I1123 20:21:26.242094 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zx5vk" podStartSLOduration=4.792768101 podStartE2EDuration="8.242079646s" podCreationTimestamp="2025-11-23 20:21:18 +0000 UTC" firstStartedPulling="2025-11-23 20:21:22.190511118 +0000 UTC m=+790.339552114" lastFinishedPulling="2025-11-23 20:21:25.639822663 +0000 UTC m=+793.788863659" observedRunningTime="2025-11-23 20:21:26.24120653 +0000 UTC m=+794.390247506" watchObservedRunningTime="2025-11-23 20:21:26.242079646 +0000 UTC m=+794.391120602" Nov 23 20:21:26 crc kubenswrapper[4726]: I1123 20:21:26.299928 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4pqht"] Nov 23 20:21:26 crc kubenswrapper[4726]: I1123 20:21:26.302759 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4pqht"] Nov 23 20:21:26 crc kubenswrapper[4726]: I1123 20:21:26.600150 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" path="/var/lib/kubelet/pods/15fc8248-951f-4041-a122-ff4f3e30050b/volumes" Nov 23 20:21:27 crc kubenswrapper[4726]: I1123 20:21:27.617275 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rq6nk"] Nov 23 20:21:27 crc kubenswrapper[4726]: I1123 20:21:27.617538 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rq6nk" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerName="registry-server" containerID="cri-o://78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2" gracePeriod=2 Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.050690 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.118836 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-catalog-content\") pod \"9338021a-a0ec-4b95-ad96-d4e95439b47b\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.118980 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmjln\" (UniqueName: \"kubernetes.io/projected/9338021a-a0ec-4b95-ad96-d4e95439b47b-kube-api-access-cmjln\") pod \"9338021a-a0ec-4b95-ad96-d4e95439b47b\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.119029 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-utilities\") pod \"9338021a-a0ec-4b95-ad96-d4e95439b47b\" (UID: \"9338021a-a0ec-4b95-ad96-d4e95439b47b\") " Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.120013 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-utilities" (OuterVolumeSpecName: "utilities") pod "9338021a-a0ec-4b95-ad96-d4e95439b47b" (UID: "9338021a-a0ec-4b95-ad96-d4e95439b47b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.126974 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9338021a-a0ec-4b95-ad96-d4e95439b47b-kube-api-access-cmjln" (OuterVolumeSpecName: "kube-api-access-cmjln") pod "9338021a-a0ec-4b95-ad96-d4e95439b47b" (UID: "9338021a-a0ec-4b95-ad96-d4e95439b47b"). InnerVolumeSpecName "kube-api-access-cmjln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.135101 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9338021a-a0ec-4b95-ad96-d4e95439b47b" (UID: "9338021a-a0ec-4b95-ad96-d4e95439b47b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.220738 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmjln\" (UniqueName: \"kubernetes.io/projected/9338021a-a0ec-4b95-ad96-d4e95439b47b-kube-api-access-cmjln\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.220781 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.220796 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9338021a-a0ec-4b95-ad96-d4e95439b47b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.264097 4726 generic.go:334] "Generic (PLEG): container finished" podID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerID="78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2" exitCode=0 Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.264143 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq6nk" event={"ID":"9338021a-a0ec-4b95-ad96-d4e95439b47b","Type":"ContainerDied","Data":"78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2"} Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.264169 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rq6nk" event={"ID":"9338021a-a0ec-4b95-ad96-d4e95439b47b","Type":"ContainerDied","Data":"bb536fa93c36ad2bbd8fe03c82c276b628eb6a42d417b0498211deb2ad577b92"} Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.264186 4726 scope.go:117] "RemoveContainer" containerID="78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.264391 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rq6nk" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.291334 4726 scope.go:117] "RemoveContainer" containerID="d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.315179 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rq6nk"] Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.324861 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rq6nk"] Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.331049 4726 scope.go:117] "RemoveContainer" containerID="236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.354478 4726 scope.go:117] "RemoveContainer" containerID="78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2" Nov 23 20:21:28 crc kubenswrapper[4726]: E1123 20:21:28.357979 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2\": container with ID starting with 78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2 not found: ID does not exist" containerID="78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.358016 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2"} err="failed to get container status \"78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2\": rpc error: code = NotFound desc = could not find container \"78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2\": container with ID starting with 78d4113c9a3fa770cd83d2d3297a4840f76bd5199219a3949398b947b273a0e2 not found: ID does not exist" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.358037 4726 scope.go:117] "RemoveContainer" containerID="d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7" Nov 23 20:21:28 crc kubenswrapper[4726]: E1123 20:21:28.358415 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7\": container with ID starting with d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7 not found: ID does not exist" containerID="d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.358460 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7"} err="failed to get container status \"d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7\": rpc error: code = NotFound desc = could not find container \"d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7\": container with ID starting with d19576ec5ebd874ff8e302ae6f49313cd05ff70a467e65dc9fc41e41a54656f7 not found: ID does not exist" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.358489 4726 scope.go:117] "RemoveContainer" containerID="236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4" Nov 23 20:21:28 crc kubenswrapper[4726]: E1123 20:21:28.358894 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4\": container with ID starting with 236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4 not found: ID does not exist" containerID="236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.358949 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4"} err="failed to get container status \"236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4\": rpc error: code = NotFound desc = could not find container \"236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4\": container with ID starting with 236af5981bde7615b517f053ba7aaa3b9d4501c1a2f7399b8d0173c1637ce1c4 not found: ID does not exist" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.567644 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.567996 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:28 crc kubenswrapper[4726]: I1123 20:21:28.599779 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" path="/var/lib/kubelet/pods/9338021a-a0ec-4b95-ad96-d4e95439b47b/volumes" Nov 23 20:21:29 crc kubenswrapper[4726]: I1123 20:21:29.611525 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zx5vk" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="registry-server" probeResult="failure" output=< Nov 23 20:21:29 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 20:21:29 crc kubenswrapper[4726]: > Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.632683 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.702898 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kd7t5"] Nov 23 20:21:38 crc kubenswrapper[4726]: E1123 20:21:38.703350 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerName="extract-content" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.703371 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerName="extract-content" Nov 23 20:21:38 crc kubenswrapper[4726]: E1123 20:21:38.703383 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" containerName="extract-content" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.703391 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" containerName="extract-content" Nov 23 20:21:38 crc kubenswrapper[4726]: E1123 20:21:38.703404 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerName="extract-utilities" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.703412 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerName="extract-utilities" Nov 23 20:21:38 crc kubenswrapper[4726]: E1123 20:21:38.703421 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerName="registry-server" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.703429 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerName="registry-server" Nov 23 20:21:38 crc kubenswrapper[4726]: E1123 20:21:38.703437 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" containerName="registry-server" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.703444 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" containerName="registry-server" Nov 23 20:21:38 crc kubenswrapper[4726]: E1123 20:21:38.703455 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" containerName="extract-utilities" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.703464 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" containerName="extract-utilities" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.703598 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="15fc8248-951f-4041-a122-ff4f3e30050b" containerName="registry-server" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.703622 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9338021a-a0ec-4b95-ad96-d4e95439b47b" containerName="registry-server" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.704755 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.725534 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kd7t5"] Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.732047 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.777551 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-utilities\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.777607 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwpzr\" (UniqueName: \"kubernetes.io/projected/6a8edd9c-fe94-4671-be59-31299be802fe-kube-api-access-zwpzr\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.777646 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-catalog-content\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.878959 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-utilities\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.879005 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwpzr\" (UniqueName: \"kubernetes.io/projected/6a8edd9c-fe94-4671-be59-31299be802fe-kube-api-access-zwpzr\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.879055 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-catalog-content\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.879526 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-catalog-content\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.879818 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-utilities\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:38 crc kubenswrapper[4726]: I1123 20:21:38.900022 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwpzr\" (UniqueName: \"kubernetes.io/projected/6a8edd9c-fe94-4671-be59-31299be802fe-kube-api-access-zwpzr\") pod \"certified-operators-kd7t5\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:39 crc kubenswrapper[4726]: I1123 20:21:39.022757 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:39 crc kubenswrapper[4726]: I1123 20:21:39.243769 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zx5vk"] Nov 23 20:21:39 crc kubenswrapper[4726]: I1123 20:21:39.587614 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kd7t5"] Nov 23 20:21:39 crc kubenswrapper[4726]: W1123 20:21:39.600168 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a8edd9c_fe94_4671_be59_31299be802fe.slice/crio-d72a628026d12caa36ff4e0639780236db26b5402a682961433cb76a132388d4 WatchSource:0}: Error finding container d72a628026d12caa36ff4e0639780236db26b5402a682961433cb76a132388d4: Status 404 returned error can't find the container with id d72a628026d12caa36ff4e0639780236db26b5402a682961433cb76a132388d4 Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.340712 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a8edd9c-fe94-4671-be59-31299be802fe" containerID="f1015cd3957770465a824c78204b4fbd3df96dab1f7ec6c3cd4ac58cba93c856" exitCode=0 Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.340825 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd7t5" event={"ID":"6a8edd9c-fe94-4671-be59-31299be802fe","Type":"ContainerDied","Data":"f1015cd3957770465a824c78204b4fbd3df96dab1f7ec6c3cd4ac58cba93c856"} Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.341100 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd7t5" event={"ID":"6a8edd9c-fe94-4671-be59-31299be802fe","Type":"ContainerStarted","Data":"d72a628026d12caa36ff4e0639780236db26b5402a682961433cb76a132388d4"} Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.341204 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zx5vk" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="registry-server" containerID="cri-o://aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37" gracePeriod=2 Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.785725 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.903165 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jkwm\" (UniqueName: \"kubernetes.io/projected/71bfcdcb-9312-4684-9b7c-3428aec8118a-kube-api-access-9jkwm\") pod \"71bfcdcb-9312-4684-9b7c-3428aec8118a\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.903299 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-catalog-content\") pod \"71bfcdcb-9312-4684-9b7c-3428aec8118a\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.903328 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-utilities\") pod \"71bfcdcb-9312-4684-9b7c-3428aec8118a\" (UID: \"71bfcdcb-9312-4684-9b7c-3428aec8118a\") " Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.904365 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-utilities" (OuterVolumeSpecName: "utilities") pod "71bfcdcb-9312-4684-9b7c-3428aec8118a" (UID: "71bfcdcb-9312-4684-9b7c-3428aec8118a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.908310 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71bfcdcb-9312-4684-9b7c-3428aec8118a-kube-api-access-9jkwm" (OuterVolumeSpecName: "kube-api-access-9jkwm") pod "71bfcdcb-9312-4684-9b7c-3428aec8118a" (UID: "71bfcdcb-9312-4684-9b7c-3428aec8118a"). InnerVolumeSpecName "kube-api-access-9jkwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:21:40 crc kubenswrapper[4726]: I1123 20:21:40.989317 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71bfcdcb-9312-4684-9b7c-3428aec8118a" (UID: "71bfcdcb-9312-4684-9b7c-3428aec8118a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.004933 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jkwm\" (UniqueName: \"kubernetes.io/projected/71bfcdcb-9312-4684-9b7c-3428aec8118a-kube-api-access-9jkwm\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.004967 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.004981 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71bfcdcb-9312-4684-9b7c-3428aec8118a-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.348279 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd7t5" event={"ID":"6a8edd9c-fe94-4671-be59-31299be802fe","Type":"ContainerStarted","Data":"1e75fbabd23d794bc28c322fb6306f4bb81e3e3dee5a3c1f797b988adddaf521"} Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.352578 4726 generic.go:334] "Generic (PLEG): container finished" podID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerID="aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37" exitCode=0 Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.352599 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx5vk" event={"ID":"71bfcdcb-9312-4684-9b7c-3428aec8118a","Type":"ContainerDied","Data":"aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37"} Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.352739 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zx5vk" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.352749 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx5vk" event={"ID":"71bfcdcb-9312-4684-9b7c-3428aec8118a","Type":"ContainerDied","Data":"12f743ceac5a658bc1aee269999f7d1706aa196b7fd8a5e592b4bc4d4ce69b43"} Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.352762 4726 scope.go:117] "RemoveContainer" containerID="aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.373108 4726 scope.go:117] "RemoveContainer" containerID="209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.406218 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zx5vk"] Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.407483 4726 scope.go:117] "RemoveContainer" containerID="8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.411321 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zx5vk"] Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.439592 4726 scope.go:117] "RemoveContainer" containerID="aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37" Nov 23 20:21:41 crc kubenswrapper[4726]: E1123 20:21:41.440374 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37\": container with ID starting with aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37 not found: ID does not exist" containerID="aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.440432 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37"} err="failed to get container status \"aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37\": rpc error: code = NotFound desc = could not find container \"aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37\": container with ID starting with aedc11b4add1c8c4efa61b007722d3f3aa8a47a113abdf9b8f21ca602f5bbd37 not found: ID does not exist" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.440464 4726 scope.go:117] "RemoveContainer" containerID="209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac" Nov 23 20:21:41 crc kubenswrapper[4726]: E1123 20:21:41.440949 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac\": container with ID starting with 209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac not found: ID does not exist" containerID="209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.441022 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac"} err="failed to get container status \"209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac\": rpc error: code = NotFound desc = could not find container \"209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac\": container with ID starting with 209c27684d892bab7ff1f1e878ab9288fbb3d0ec6e2ed4d08d453fc9d12744ac not found: ID does not exist" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.441058 4726 scope.go:117] "RemoveContainer" containerID="8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7" Nov 23 20:21:41 crc kubenswrapper[4726]: E1123 20:21:41.441534 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7\": container with ID starting with 8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7 not found: ID does not exist" containerID="8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7" Nov 23 20:21:41 crc kubenswrapper[4726]: I1123 20:21:41.441560 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7"} err="failed to get container status \"8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7\": rpc error: code = NotFound desc = could not find container \"8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7\": container with ID starting with 8a8817a2993efd3c01399da681621361a16792be31f72657c35d324fdec6dce7 not found: ID does not exist" Nov 23 20:21:42 crc kubenswrapper[4726]: I1123 20:21:42.366900 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a8edd9c-fe94-4671-be59-31299be802fe" containerID="1e75fbabd23d794bc28c322fb6306f4bb81e3e3dee5a3c1f797b988adddaf521" exitCode=0 Nov 23 20:21:42 crc kubenswrapper[4726]: I1123 20:21:42.366949 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd7t5" event={"ID":"6a8edd9c-fe94-4671-be59-31299be802fe","Type":"ContainerDied","Data":"1e75fbabd23d794bc28c322fb6306f4bb81e3e3dee5a3c1f797b988adddaf521"} Nov 23 20:21:42 crc kubenswrapper[4726]: I1123 20:21:42.601455 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" path="/var/lib/kubelet/pods/71bfcdcb-9312-4684-9b7c-3428aec8118a/volumes" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.373380 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd7t5" event={"ID":"6a8edd9c-fe94-4671-be59-31299be802fe","Type":"ContainerStarted","Data":"443f2466af374ec6ab66aba4901ffb04089b458155fc45edc59498224429efa5"} Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.400809 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kd7t5" podStartSLOduration=2.915778068 podStartE2EDuration="5.400792269s" podCreationTimestamp="2025-11-23 20:21:38 +0000 UTC" firstStartedPulling="2025-11-23 20:21:40.342307342 +0000 UTC m=+808.491348288" lastFinishedPulling="2025-11-23 20:21:42.827321543 +0000 UTC m=+810.976362489" observedRunningTime="2025-11-23 20:21:43.397004809 +0000 UTC m=+811.546045765" watchObservedRunningTime="2025-11-23 20:21:43.400792269 +0000 UTC m=+811.549833225" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.471560 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp"] Nov 23 20:21:43 crc kubenswrapper[4726]: E1123 20:21:43.471834 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="extract-utilities" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.471852 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="extract-utilities" Nov 23 20:21:43 crc kubenswrapper[4726]: E1123 20:21:43.471887 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="extract-content" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.471894 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="extract-content" Nov 23 20:21:43 crc kubenswrapper[4726]: E1123 20:21:43.471905 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="registry-server" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.471911 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="registry-server" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.472030 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="71bfcdcb-9312-4684-9b7c-3428aec8118a" containerName="registry-server" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.472597 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.474596 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hnzv7" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.474948 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.475877 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.480534 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-j8qzg" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.510953 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.515628 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.516529 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.518809 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-mtvw4" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.537938 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqx2l\" (UniqueName: \"kubernetes.io/projected/d481bfb6-f28c-49ed-8efe-7fc35dc6d608-kube-api-access-dqx2l\") pod \"barbican-operator-controller-manager-7768f8c84f-q7qxp\" (UID: \"d481bfb6-f28c-49ed-8efe-7fc35dc6d608\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.538019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7smb\" (UniqueName: \"kubernetes.io/projected/a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7-kube-api-access-m7smb\") pod \"cinder-operator-controller-manager-6d8fd67bf7-xt9t6\" (UID: \"a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.539238 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.556941 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.563152 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.564001 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.567005 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-rvh5v" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.587289 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.588140 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.604591 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-zcpmj" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.626035 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.639584 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7smb\" (UniqueName: \"kubernetes.io/projected/a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7-kube-api-access-m7smb\") pod \"cinder-operator-controller-manager-6d8fd67bf7-xt9t6\" (UID: \"a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.639647 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czcsq\" (UniqueName: \"kubernetes.io/projected/2e4ac3d0-65ba-4140-9a03-c269dfebad67-kube-api-access-czcsq\") pod \"designate-operator-controller-manager-56dfb6b67f-wh58z\" (UID: \"2e4ac3d0-65ba-4140-9a03-c269dfebad67\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.639712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqx2l\" (UniqueName: \"kubernetes.io/projected/d481bfb6-f28c-49ed-8efe-7fc35dc6d608-kube-api-access-dqx2l\") pod \"barbican-operator-controller-manager-7768f8c84f-q7qxp\" (UID: \"d481bfb6-f28c-49ed-8efe-7fc35dc6d608\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.639754 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg2w4\" (UniqueName: \"kubernetes.io/projected/9c05c149-b530-4ca3-b201-309826c5e095-kube-api-access-cg2w4\") pod \"heat-operator-controller-manager-bf4c6585d-jc55l\" (UID: \"9c05c149-b530-4ca3-b201-309826c5e095\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.639790 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trmds\" (UniqueName: \"kubernetes.io/projected/bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb-kube-api-access-trmds\") pod \"glance-operator-controller-manager-8667fbf6f6-6d8lj\" (UID: \"bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.642212 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.651983 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.653056 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.658205 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-lmqld" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.661764 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.667275 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7smb\" (UniqueName: \"kubernetes.io/projected/a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7-kube-api-access-m7smb\") pod \"cinder-operator-controller-manager-6d8fd67bf7-xt9t6\" (UID: \"a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.672648 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.673565 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.694054 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.697570 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.705575 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-q24vt" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.721945 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqx2l\" (UniqueName: \"kubernetes.io/projected/d481bfb6-f28c-49ed-8efe-7fc35dc6d608-kube-api-access-dqx2l\") pod \"barbican-operator-controller-manager-7768f8c84f-q7qxp\" (UID: \"d481bfb6-f28c-49ed-8efe-7fc35dc6d608\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.751854 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89pgj\" (UniqueName: \"kubernetes.io/projected/c3279a0f-e448-47b2-91b3-6de329d77cec-kube-api-access-89pgj\") pod \"horizon-operator-controller-manager-5d86b44686-574zb\" (UID: \"c3279a0f-e448-47b2-91b3-6de329d77cec\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.752004 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vhvn\" (UniqueName: \"kubernetes.io/projected/765e2946-d293-4840-baf2-96a478748a25-kube-api-access-5vhvn\") pod \"infra-operator-controller-manager-769d9c7585-w4lxk\" (UID: \"765e2946-d293-4840-baf2-96a478748a25\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.752093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg2w4\" (UniqueName: \"kubernetes.io/projected/9c05c149-b530-4ca3-b201-309826c5e095-kube-api-access-cg2w4\") pod \"heat-operator-controller-manager-bf4c6585d-jc55l\" (UID: \"9c05c149-b530-4ca3-b201-309826c5e095\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.752135 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/765e2946-d293-4840-baf2-96a478748a25-cert\") pod \"infra-operator-controller-manager-769d9c7585-w4lxk\" (UID: \"765e2946-d293-4840-baf2-96a478748a25\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.752235 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trmds\" (UniqueName: \"kubernetes.io/projected/bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb-kube-api-access-trmds\") pod \"glance-operator-controller-manager-8667fbf6f6-6d8lj\" (UID: \"bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.752360 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czcsq\" (UniqueName: \"kubernetes.io/projected/2e4ac3d0-65ba-4140-9a03-c269dfebad67-kube-api-access-czcsq\") pod \"designate-operator-controller-manager-56dfb6b67f-wh58z\" (UID: \"2e4ac3d0-65ba-4140-9a03-c269dfebad67\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.789035 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.798638 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.817493 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czcsq\" (UniqueName: \"kubernetes.io/projected/2e4ac3d0-65ba-4140-9a03-c269dfebad67-kube-api-access-czcsq\") pod \"designate-operator-controller-manager-56dfb6b67f-wh58z\" (UID: \"2e4ac3d0-65ba-4140-9a03-c269dfebad67\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.817894 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.819733 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.820806 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trmds\" (UniqueName: \"kubernetes.io/projected/bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb-kube-api-access-trmds\") pod \"glance-operator-controller-manager-8667fbf6f6-6d8lj\" (UID: \"bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.826967 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-pcnt8" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.839999 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.873518 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg2w4\" (UniqueName: \"kubernetes.io/projected/9c05c149-b530-4ca3-b201-309826c5e095-kube-api-access-cg2w4\") pod \"heat-operator-controller-manager-bf4c6585d-jc55l\" (UID: \"9c05c149-b530-4ca3-b201-309826c5e095\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.874133 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89pgj\" (UniqueName: \"kubernetes.io/projected/c3279a0f-e448-47b2-91b3-6de329d77cec-kube-api-access-89pgj\") pod \"horizon-operator-controller-manager-5d86b44686-574zb\" (UID: \"c3279a0f-e448-47b2-91b3-6de329d77cec\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.874160 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vhvn\" (UniqueName: \"kubernetes.io/projected/765e2946-d293-4840-baf2-96a478748a25-kube-api-access-5vhvn\") pod \"infra-operator-controller-manager-769d9c7585-w4lxk\" (UID: \"765e2946-d293-4840-baf2-96a478748a25\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.874184 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/765e2946-d293-4840-baf2-96a478748a25-cert\") pod \"infra-operator-controller-manager-769d9c7585-w4lxk\" (UID: \"765e2946-d293-4840-baf2-96a478748a25\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:43 crc kubenswrapper[4726]: E1123 20:21:43.874296 4726 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 23 20:21:43 crc kubenswrapper[4726]: E1123 20:21:43.874334 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765e2946-d293-4840-baf2-96a478748a25-cert podName:765e2946-d293-4840-baf2-96a478748a25 nodeName:}" failed. No retries permitted until 2025-11-23 20:21:44.374320696 +0000 UTC m=+812.523361652 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/765e2946-d293-4840-baf2-96a478748a25-cert") pod "infra-operator-controller-manager-769d9c7585-w4lxk" (UID: "765e2946-d293-4840-baf2-96a478748a25") : secret "infra-operator-webhook-server-cert" not found Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.876718 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.877250 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.877972 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.891616 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-tmshx" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.904018 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.905514 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.908610 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-lhj24" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.914554 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89pgj\" (UniqueName: \"kubernetes.io/projected/c3279a0f-e448-47b2-91b3-6de329d77cec-kube-api-access-89pgj\") pod \"horizon-operator-controller-manager-5d86b44686-574zb\" (UID: \"c3279a0f-e448-47b2-91b3-6de329d77cec\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.914618 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vhvn\" (UniqueName: \"kubernetes.io/projected/765e2946-d293-4840-baf2-96a478748a25-kube-api-access-5vhvn\") pod \"infra-operator-controller-manager-769d9c7585-w4lxk\" (UID: \"765e2946-d293-4840-baf2-96a478748a25\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.929200 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.932369 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.959371 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.973916 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4"] Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.975560 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhgh2\" (UniqueName: \"kubernetes.io/projected/a653d1f5-fe23-4a55-9422-33964eeb8488-kube-api-access-zhgh2\") pod \"manila-operator-controller-manager-7bb88cb858-5s9v2\" (UID: \"a653d1f5-fe23-4a55-9422-33964eeb8488\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.975697 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgzbt\" (UniqueName: \"kubernetes.io/projected/76e7ee34-d8e7-4d5b-b79e-acd11b2b0735-kube-api-access-kgzbt\") pod \"keystone-operator-controller-manager-7879fb76fd-s7746\" (UID: \"76e7ee34-d8e7-4d5b-b79e-acd11b2b0735\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" Nov 23 20:21:43 crc kubenswrapper[4726]: I1123 20:21:43.975772 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z64r2\" (UniqueName: \"kubernetes.io/projected/c53c8659-c283-4c78-88b4-b60f8b040dd7-kube-api-access-z64r2\") pod \"ironic-operator-controller-manager-5c75d7c94b-m4xc4\" (UID: \"c53c8659-c283-4c78-88b4-b60f8b040dd7\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.008921 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.022111 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.024662 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.052275 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-jh6jw" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.079352 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgzbt\" (UniqueName: \"kubernetes.io/projected/76e7ee34-d8e7-4d5b-b79e-acd11b2b0735-kube-api-access-kgzbt\") pod \"keystone-operator-controller-manager-7879fb76fd-s7746\" (UID: \"76e7ee34-d8e7-4d5b-b79e-acd11b2b0735\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.079380 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z64r2\" (UniqueName: \"kubernetes.io/projected/c53c8659-c283-4c78-88b4-b60f8b040dd7-kube-api-access-z64r2\") pod \"ironic-operator-controller-manager-5c75d7c94b-m4xc4\" (UID: \"c53c8659-c283-4c78-88b4-b60f8b040dd7\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.079467 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhgh2\" (UniqueName: \"kubernetes.io/projected/a653d1f5-fe23-4a55-9422-33964eeb8488-kube-api-access-zhgh2\") pod \"manila-operator-controller-manager-7bb88cb858-5s9v2\" (UID: \"a653d1f5-fe23-4a55-9422-33964eeb8488\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.079504 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdmbz\" (UniqueName: \"kubernetes.io/projected/b89cb12d-b390-40ed-908f-e1095fe0c62b-kube-api-access-xdmbz\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-8dz5d\" (UID: \"b89cb12d-b390-40ed-908f-e1095fe0c62b\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.085918 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.122085 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z64r2\" (UniqueName: \"kubernetes.io/projected/c53c8659-c283-4c78-88b4-b60f8b040dd7-kube-api-access-z64r2\") pod \"ironic-operator-controller-manager-5c75d7c94b-m4xc4\" (UID: \"c53c8659-c283-4c78-88b4-b60f8b040dd7\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.132327 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhgh2\" (UniqueName: \"kubernetes.io/projected/a653d1f5-fe23-4a55-9422-33964eeb8488-kube-api-access-zhgh2\") pod \"manila-operator-controller-manager-7bb88cb858-5s9v2\" (UID: \"a653d1f5-fe23-4a55-9422-33964eeb8488\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.133811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgzbt\" (UniqueName: \"kubernetes.io/projected/76e7ee34-d8e7-4d5b-b79e-acd11b2b0735-kube-api-access-kgzbt\") pod \"keystone-operator-controller-manager-7879fb76fd-s7746\" (UID: \"76e7ee34-d8e7-4d5b-b79e-acd11b2b0735\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.142983 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.145473 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.155469 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-nmkmk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.165716 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.166979 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.172348 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-6drzv" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.184272 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdmbz\" (UniqueName: \"kubernetes.io/projected/b89cb12d-b390-40ed-908f-e1095fe0c62b-kube-api-access-xdmbz\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-8dz5d\" (UID: \"b89cb12d-b390-40ed-908f-e1095fe0c62b\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.184325 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6npmp\" (UniqueName: \"kubernetes.io/projected/4301d05b-4769-4c53-b557-bd6aa37efa68-kube-api-access-6npmp\") pod \"neutron-operator-controller-manager-66b7d6f598-4p7kg\" (UID: \"4301d05b-4769-4c53-b557-bd6aa37efa68\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.230524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdmbz\" (UniqueName: \"kubernetes.io/projected/b89cb12d-b390-40ed-908f-e1095fe0c62b-kube-api-access-xdmbz\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-8dz5d\" (UID: \"b89cb12d-b390-40ed-908f-e1095fe0c62b\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.256308 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.260727 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.276181 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.292128 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqrsd\" (UniqueName: \"kubernetes.io/projected/384a1b22-97f2-4f0b-ad98-52adbbf5178b-kube-api-access-cqrsd\") pod \"nova-operator-controller-manager-86d796d84d-qvx6m\" (UID: \"384a1b22-97f2-4f0b-ad98-52adbbf5178b\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.292227 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6npmp\" (UniqueName: \"kubernetes.io/projected/4301d05b-4769-4c53-b557-bd6aa37efa68-kube-api-access-6npmp\") pod \"neutron-operator-controller-manager-66b7d6f598-4p7kg\" (UID: \"4301d05b-4769-4c53-b557-bd6aa37efa68\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.306470 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.309841 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.326638 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.326820 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-946m8" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.328567 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.338972 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6npmp\" (UniqueName: \"kubernetes.io/projected/4301d05b-4769-4c53-b557-bd6aa37efa68-kube-api-access-6npmp\") pod \"neutron-operator-controller-manager-66b7d6f598-4p7kg\" (UID: \"4301d05b-4769-4c53-b557-bd6aa37efa68\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.340052 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.341108 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.348328 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-vw9bl" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.353918 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.363757 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.379665 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.401912 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqrsd\" (UniqueName: \"kubernetes.io/projected/384a1b22-97f2-4f0b-ad98-52adbbf5178b-kube-api-access-cqrsd\") pod \"nova-operator-controller-manager-86d796d84d-qvx6m\" (UID: \"384a1b22-97f2-4f0b-ad98-52adbbf5178b\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.401998 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/765e2946-d293-4840-baf2-96a478748a25-cert\") pod \"infra-operator-controller-manager-769d9c7585-w4lxk\" (UID: \"765e2946-d293-4840-baf2-96a478748a25\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.402029 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44hnx49\" (UID: \"ce7ed28d-d56b-44de-b207-7c47ab684921\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.402066 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkjqv\" (UniqueName: \"kubernetes.io/projected/ce7ed28d-d56b-44de-b207-7c47ab684921-kube-api-access-qkjqv\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44hnx49\" (UID: \"ce7ed28d-d56b-44de-b207-7c47ab684921\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.402102 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvsdd\" (UniqueName: \"kubernetes.io/projected/20b0ca2b-a3ee-45c9-8b34-b476245a4ed6-kube-api-access-mvsdd\") pod \"octavia-operator-controller-manager-6fdc856c5d-7lwwg\" (UID: \"20b0ca2b-a3ee-45c9-8b34-b476245a4ed6\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.412658 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.412988 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/765e2946-d293-4840-baf2-96a478748a25-cert\") pod \"infra-operator-controller-manager-769d9c7585-w4lxk\" (UID: \"765e2946-d293-4840-baf2-96a478748a25\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.413083 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.414079 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.422104 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-vbc5z" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.428668 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.429741 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.433171 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-tp64h" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.439605 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.443491 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqrsd\" (UniqueName: \"kubernetes.io/projected/384a1b22-97f2-4f0b-ad98-52adbbf5178b-kube-api-access-cqrsd\") pod \"nova-operator-controller-manager-86d796d84d-qvx6m\" (UID: \"384a1b22-97f2-4f0b-ad98-52adbbf5178b\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.451067 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.451995 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.464231 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-s2lk5" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.492776 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.507508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44hnx49\" (UID: \"ce7ed28d-d56b-44de-b207-7c47ab684921\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.507558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkjqv\" (UniqueName: \"kubernetes.io/projected/ce7ed28d-d56b-44de-b207-7c47ab684921-kube-api-access-qkjqv\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44hnx49\" (UID: \"ce7ed28d-d56b-44de-b207-7c47ab684921\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.507590 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnbp8\" (UniqueName: \"kubernetes.io/projected/cfdab430-7dd4-4612-a65b-b381b5b58115-kube-api-access-tnbp8\") pod \"swift-operator-controller-manager-799cb6ffd6-frss6\" (UID: \"cfdab430-7dd4-4612-a65b-b381b5b58115\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.507607 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvsdd\" (UniqueName: \"kubernetes.io/projected/20b0ca2b-a3ee-45c9-8b34-b476245a4ed6-kube-api-access-mvsdd\") pod \"octavia-operator-controller-manager-6fdc856c5d-7lwwg\" (UID: \"20b0ca2b-a3ee-45c9-8b34-b476245a4ed6\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.507637 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl5xh\" (UniqueName: \"kubernetes.io/projected/42d03b10-01f7-4963-a97b-f3d443a1cdf7-kube-api-access-hl5xh\") pod \"placement-operator-controller-manager-6dc664666c-p7r99\" (UID: \"42d03b10-01f7-4963-a97b-f3d443a1cdf7\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.507657 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x4vn\" (UniqueName: \"kubernetes.io/projected/6e692743-eb35-4a91-97af-8d86cf694470-kube-api-access-8x4vn\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-2k5zx\" (UID: \"6e692743-eb35-4a91-97af-8d86cf694470\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" Nov 23 20:21:44 crc kubenswrapper[4726]: E1123 20:21:44.511379 4726 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 23 20:21:44 crc kubenswrapper[4726]: E1123 20:21:44.511433 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert podName:ce7ed28d-d56b-44de-b207-7c47ab684921 nodeName:}" failed. No retries permitted until 2025-11-23 20:21:45.011414782 +0000 UTC m=+813.160455738 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert") pod "openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" (UID: "ce7ed28d-d56b-44de-b207-7c47ab684921") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.526569 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.526658 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.540519 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.541609 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.547383 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-hbd4x" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.550336 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.581324 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvsdd\" (UniqueName: \"kubernetes.io/projected/20b0ca2b-a3ee-45c9-8b34-b476245a4ed6-kube-api-access-mvsdd\") pod \"octavia-operator-controller-manager-6fdc856c5d-7lwwg\" (UID: \"20b0ca2b-a3ee-45c9-8b34-b476245a4ed6\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.630369 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkjqv\" (UniqueName: \"kubernetes.io/projected/ce7ed28d-d56b-44de-b207-7c47ab684921-kube-api-access-qkjqv\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44hnx49\" (UID: \"ce7ed28d-d56b-44de-b207-7c47ab684921\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.636525 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl5xh\" (UniqueName: \"kubernetes.io/projected/42d03b10-01f7-4963-a97b-f3d443a1cdf7-kube-api-access-hl5xh\") pod \"placement-operator-controller-manager-6dc664666c-p7r99\" (UID: \"42d03b10-01f7-4963-a97b-f3d443a1cdf7\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.636571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq8pc\" (UniqueName: \"kubernetes.io/projected/74d488f0-9a23-4774-94e7-75605573ce43-kube-api-access-lq8pc\") pod \"telemetry-operator-controller-manager-7798859c74-527bk\" (UID: \"74d488f0-9a23-4774-94e7-75605573ce43\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.636602 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x4vn\" (UniqueName: \"kubernetes.io/projected/6e692743-eb35-4a91-97af-8d86cf694470-kube-api-access-8x4vn\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-2k5zx\" (UID: \"6e692743-eb35-4a91-97af-8d86cf694470\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.636717 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnbp8\" (UniqueName: \"kubernetes.io/projected/cfdab430-7dd4-4612-a65b-b381b5b58115-kube-api-access-tnbp8\") pod \"swift-operator-controller-manager-799cb6ffd6-frss6\" (UID: \"cfdab430-7dd4-4612-a65b-b381b5b58115\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.644486 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.665780 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.666070 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.676350 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.680786 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-b86s2" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.690034 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.694577 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.706668 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-vdxq6" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.748880 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq8pc\" (UniqueName: \"kubernetes.io/projected/74d488f0-9a23-4774-94e7-75605573ce43-kube-api-access-lq8pc\") pod \"telemetry-operator-controller-manager-7798859c74-527bk\" (UID: \"74d488f0-9a23-4774-94e7-75605573ce43\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.748951 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b84np\" (UniqueName: \"kubernetes.io/projected/023e8819-c436-4592-a6f6-ea3448733363-kube-api-access-b84np\") pod \"watcher-operator-controller-manager-7cd4fb6f79-7x7b7\" (UID: \"023e8819-c436-4592-a6f6-ea3448733363\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.756102 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x4vn\" (UniqueName: \"kubernetes.io/projected/6e692743-eb35-4a91-97af-8d86cf694470-kube-api-access-8x4vn\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-2k5zx\" (UID: \"6e692743-eb35-4a91-97af-8d86cf694470\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.757038 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnbp8\" (UniqueName: \"kubernetes.io/projected/cfdab430-7dd4-4612-a65b-b381b5b58115-kube-api-access-tnbp8\") pod \"swift-operator-controller-manager-799cb6ffd6-frss6\" (UID: \"cfdab430-7dd4-4612-a65b-b381b5b58115\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.793609 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl5xh\" (UniqueName: \"kubernetes.io/projected/42d03b10-01f7-4963-a97b-f3d443a1cdf7-kube-api-access-hl5xh\") pod \"placement-operator-controller-manager-6dc664666c-p7r99\" (UID: \"42d03b10-01f7-4963-a97b-f3d443a1cdf7\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.798384 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.838377 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.840756 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.845383 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq8pc\" (UniqueName: \"kubernetes.io/projected/74d488f0-9a23-4774-94e7-75605573ce43-kube-api-access-lq8pc\") pod \"telemetry-operator-controller-manager-7798859c74-527bk\" (UID: \"74d488f0-9a23-4774-94e7-75605573ce43\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.850157 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79t84\" (UniqueName: \"kubernetes.io/projected/c22e9617-abf8-4370-ba10-31e967257270-kube-api-access-79t84\") pod \"test-operator-controller-manager-8464cf66df-5w2nt\" (UID: \"c22e9617-abf8-4370-ba10-31e967257270\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.850199 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b84np\" (UniqueName: \"kubernetes.io/projected/023e8819-c436-4592-a6f6-ea3448733363-kube-api-access-b84np\") pod \"watcher-operator-controller-manager-7cd4fb6f79-7x7b7\" (UID: \"023e8819-c436-4592-a6f6-ea3448733363\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.860419 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.866976 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.884589 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.895512 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt"] Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.904682 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b84np\" (UniqueName: \"kubernetes.io/projected/023e8819-c436-4592-a6f6-ea3448733363-kube-api-access-b84np\") pod \"watcher-operator-controller-manager-7cd4fb6f79-7x7b7\" (UID: \"023e8819-c436-4592-a6f6-ea3448733363\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" Nov 23 20:21:44 crc kubenswrapper[4726]: I1123 20:21:44.951683 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79t84\" (UniqueName: \"kubernetes.io/projected/c22e9617-abf8-4370-ba10-31e967257270-kube-api-access-79t84\") pod \"test-operator-controller-manager-8464cf66df-5w2nt\" (UID: \"c22e9617-abf8-4370-ba10-31e967257270\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.001520 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk"] Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.003379 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.032512 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79t84\" (UniqueName: \"kubernetes.io/projected/c22e9617-abf8-4370-ba10-31e967257270-kube-api-access-79t84\") pod \"test-operator-controller-manager-8464cf66df-5w2nt\" (UID: \"c22e9617-abf8-4370-ba10-31e967257270\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.032984 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-vkmfb" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.033769 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.059639 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44hnx49\" (UID: \"ce7ed28d-d56b-44de-b207-7c47ab684921\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:45 crc kubenswrapper[4726]: E1123 20:21:45.059818 4726 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 23 20:21:45 crc kubenswrapper[4726]: E1123 20:21:45.059893 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert podName:ce7ed28d-d56b-44de-b207-7c47ab684921 nodeName:}" failed. No retries permitted until 2025-11-23 20:21:46.059851667 +0000 UTC m=+814.208892623 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert") pod "openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" (UID: "ce7ed28d-d56b-44de-b207-7c47ab684921") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.082950 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk"] Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.094508 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.148848 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.160107 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv"] Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.160996 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.165631 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5g2w\" (UniqueName: \"kubernetes.io/projected/706d33da-dac2-4403-a4d8-d8e7f04f772b-kube-api-access-x5g2w\") pod \"openstack-operator-controller-manager-6cb9dc54f8-t9ptk\" (UID: \"706d33da-dac2-4403-a4d8-d8e7f04f772b\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.165686 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706d33da-dac2-4403-a4d8-d8e7f04f772b-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-t9ptk\" (UID: \"706d33da-dac2-4403-a4d8-d8e7f04f772b\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.168159 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-66jd9" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.179402 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv"] Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.206691 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6"] Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.247247 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp"] Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.266475 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5g2w\" (UniqueName: \"kubernetes.io/projected/706d33da-dac2-4403-a4d8-d8e7f04f772b-kube-api-access-x5g2w\") pod \"openstack-operator-controller-manager-6cb9dc54f8-t9ptk\" (UID: \"706d33da-dac2-4403-a4d8-d8e7f04f772b\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.266537 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706d33da-dac2-4403-a4d8-d8e7f04f772b-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-t9ptk\" (UID: \"706d33da-dac2-4403-a4d8-d8e7f04f772b\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.266562 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tztkl\" (UniqueName: \"kubernetes.io/projected/e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9-kube-api-access-tztkl\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv\" (UID: \"e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" Nov 23 20:21:45 crc kubenswrapper[4726]: E1123 20:21:45.266826 4726 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 23 20:21:45 crc kubenswrapper[4726]: E1123 20:21:45.266925 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/706d33da-dac2-4403-a4d8-d8e7f04f772b-cert podName:706d33da-dac2-4403-a4d8-d8e7f04f772b nodeName:}" failed. No retries permitted until 2025-11-23 20:21:45.766901866 +0000 UTC m=+813.915942822 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/706d33da-dac2-4403-a4d8-d8e7f04f772b-cert") pod "openstack-operator-controller-manager-6cb9dc54f8-t9ptk" (UID: "706d33da-dac2-4403-a4d8-d8e7f04f772b") : secret "webhook-server-cert" not found Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.325387 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5g2w\" (UniqueName: \"kubernetes.io/projected/706d33da-dac2-4403-a4d8-d8e7f04f772b-kube-api-access-x5g2w\") pod \"openstack-operator-controller-manager-6cb9dc54f8-t9ptk\" (UID: \"706d33da-dac2-4403-a4d8-d8e7f04f772b\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.385538 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tztkl\" (UniqueName: \"kubernetes.io/projected/e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9-kube-api-access-tztkl\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv\" (UID: \"e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.427468 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tztkl\" (UniqueName: \"kubernetes.io/projected/e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9-kube-api-access-tztkl\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv\" (UID: \"e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.458091 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" event={"ID":"d481bfb6-f28c-49ed-8efe-7fc35dc6d608","Type":"ContainerStarted","Data":"1af25fd23259924fe9a3a0fff7c906348bca121d1ee39aa24ea154a34db56e37"} Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.481990 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" event={"ID":"a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7","Type":"ContainerStarted","Data":"4d5e8e084d7bfc29143291adb46be44315d8c91983ee17aee3a5f83919deafc2"} Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.488028 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.533776 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z"] Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.794410 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706d33da-dac2-4403-a4d8-d8e7f04f772b-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-t9ptk\" (UID: \"706d33da-dac2-4403-a4d8-d8e7f04f772b\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.804371 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/706d33da-dac2-4403-a4d8-d8e7f04f772b-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-t9ptk\" (UID: \"706d33da-dac2-4403-a4d8-d8e7f04f772b\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.934135 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj"] Nov 23 20:21:45 crc kubenswrapper[4726]: I1123 20:21:45.952191 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4"] Nov 23 20:21:45 crc kubenswrapper[4726]: W1123 20:21:45.953793 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf353aea_5f6e_4aa1_9a83_ac4e9b3edbfb.slice/crio-f9241a32276cbcac969e9ffa28cd7935f1625ffa165fe0839ab74f32bdb34061 WatchSource:0}: Error finding container f9241a32276cbcac969e9ffa28cd7935f1625ffa165fe0839ab74f32bdb34061: Status 404 returned error can't find the container with id f9241a32276cbcac969e9ffa28cd7935f1625ffa165fe0839ab74f32bdb34061 Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.008455 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.024013 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.039379 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.101370 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44hnx49\" (UID: \"ce7ed28d-d56b-44de-b207-7c47ab684921\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.108229 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce7ed28d-d56b-44de-b207-7c47ab684921-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44hnx49\" (UID: \"ce7ed28d-d56b-44de-b207-7c47ab684921\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.162150 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.509976 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" event={"ID":"9c05c149-b530-4ca3-b201-309826c5e095","Type":"ContainerStarted","Data":"8f9fbe581d77025ba2032a9e5381aa02cee970a56c8c106a18b3ad1970ae42cb"} Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.522740 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" event={"ID":"c3279a0f-e448-47b2-91b3-6de329d77cec","Type":"ContainerStarted","Data":"e730d2519cce01369edff64d4347a7f2a6aad3a411b484b6b8663d413afe4ccc"} Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.532021 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" event={"ID":"c53c8659-c283-4c78-88b4-b60f8b040dd7","Type":"ContainerStarted","Data":"f252c1e03fb59c3842cbb6be0b628201145416275163a5ca8ff05778c873261a"} Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.537840 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" event={"ID":"bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb","Type":"ContainerStarted","Data":"f9241a32276cbcac969e9ffa28cd7935f1625ffa165fe0839ab74f32bdb34061"} Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.543696 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" event={"ID":"2e4ac3d0-65ba-4140-9a03-c269dfebad67","Type":"ContainerStarted","Data":"159b01cd1d01b1949444adcca34275b223e23fe2f5e276729c235e84b0f47336"} Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.718677 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.728231 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.737944 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.755947 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.835299 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.866678 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.878663 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.883352 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.887920 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.893194 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.896910 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m"] Nov 23 20:21:46 crc kubenswrapper[4726]: I1123 20:21:46.902918 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg"] Nov 23 20:21:46 crc kubenswrapper[4726]: E1123 20:21:46.936981 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6npmp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-66b7d6f598-4p7kg_openstack-operators(4301d05b-4769-4c53-b557-bd6aa37efa68): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 20:21:46 crc kubenswrapper[4726]: W1123 20:21:46.972053 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod384a1b22_97f2_4f0b_ad98_52adbbf5178b.slice/crio-32681969e0c0459f230e00e0aa9f2e9c135f1d7bfb52cdc9522603307cf600ec WatchSource:0}: Error finding container 32681969e0c0459f230e00e0aa9f2e9c135f1d7bfb52cdc9522603307cf600ec: Status 404 returned error can't find the container with id 32681969e0c0459f230e00e0aa9f2e9c135f1d7bfb52cdc9522603307cf600ec Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.071228 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk"] Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.074557 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv"] Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.099640 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt"] Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.127112 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49"] Nov 23 20:21:47 crc kubenswrapper[4726]: E1123 20:21:47.127176 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tztkl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv_openstack-operators(e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 20:21:47 crc kubenswrapper[4726]: E1123 20:21:47.130820 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" podUID="e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9" Nov 23 20:21:47 crc kubenswrapper[4726]: W1123 20:21:47.131877 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc22e9617_abf8_4370_ba10_31e967257270.slice/crio-274e90aff18808a5e40bdf3cb85f759bf084920531ba00a20cf423185186ba21 WatchSource:0}: Error finding container 274e90aff18808a5e40bdf3cb85f759bf084920531ba00a20cf423185186ba21: Status 404 returned error can't find the container with id 274e90aff18808a5e40bdf3cb85f759bf084920531ba00a20cf423185186ba21 Nov 23 20:21:47 crc kubenswrapper[4726]: E1123 20:21:47.157461 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-79t84,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-8464cf66df-5w2nt_openstack-operators(c22e9617-abf8-4370-ba10-31e967257270): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 23 20:21:47 crc kubenswrapper[4726]: E1123 20:21:47.240139 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" podUID="4301d05b-4769-4c53-b557-bd6aa37efa68" Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.569419 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" event={"ID":"c22e9617-abf8-4370-ba10-31e967257270","Type":"ContainerStarted","Data":"274e90aff18808a5e40bdf3cb85f759bf084920531ba00a20cf423185186ba21"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.572088 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" event={"ID":"42d03b10-01f7-4963-a97b-f3d443a1cdf7","Type":"ContainerStarted","Data":"28a4cddf605ebda577986d4ee63290ed4c47b1ac276ee93a1119f8e203aa528a"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.573254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" event={"ID":"20b0ca2b-a3ee-45c9-8b34-b476245a4ed6","Type":"ContainerStarted","Data":"f708d3cd533d96be6628023a83144526d7623f50e698933bafeed567f6b9b08e"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.581517 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" event={"ID":"706d33da-dac2-4403-a4d8-d8e7f04f772b","Type":"ContainerStarted","Data":"e502abbbb748a4d89e40a4de27e8074f697eb9b125a203c51402f69b0b4131cb"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.584336 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" event={"ID":"ce7ed28d-d56b-44de-b207-7c47ab684921","Type":"ContainerStarted","Data":"4d959d5de16f237e93eec4a7c9009bc270a5f92a72f10e4623f5ea3ac88cdaa3"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.585765 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" event={"ID":"a653d1f5-fe23-4a55-9422-33964eeb8488","Type":"ContainerStarted","Data":"8a860706ab9e2613932776084bc586ca59e7ec0349efde4faaf96649fee6d6b9"} Nov 23 20:21:47 crc kubenswrapper[4726]: E1123 20:21:47.592904 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" podUID="c22e9617-abf8-4370-ba10-31e967257270" Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.593267 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" event={"ID":"76e7ee34-d8e7-4d5b-b79e-acd11b2b0735","Type":"ContainerStarted","Data":"7f829d02005249e72b19e0e723667f8bc8010c974b1687bf52969cd79e006851"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.603158 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" event={"ID":"765e2946-d293-4840-baf2-96a478748a25","Type":"ContainerStarted","Data":"334b84ad4a0fce4ffead66bed2224e0170a6b7f24d63a441dd6bfc90813a8a42"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.626208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" event={"ID":"b89cb12d-b390-40ed-908f-e1095fe0c62b","Type":"ContainerStarted","Data":"75c34ee4049272ad112e06662b408eff1a9f69d339ca11eda11ea93c42e80cb0"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.638664 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" event={"ID":"74d488f0-9a23-4774-94e7-75605573ce43","Type":"ContainerStarted","Data":"00282afb1b279ee6627463be229b9e6e76d6c88da46b4a91f3e735b7ae243c3d"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.659161 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" event={"ID":"6e692743-eb35-4a91-97af-8d86cf694470","Type":"ContainerStarted","Data":"c1335eaf450be612621bbda35c79275bcb7359eaabc5964a671c6aaa671ba3b1"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.666089 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" event={"ID":"384a1b22-97f2-4f0b-ad98-52adbbf5178b","Type":"ContainerStarted","Data":"32681969e0c0459f230e00e0aa9f2e9c135f1d7bfb52cdc9522603307cf600ec"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.667537 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" event={"ID":"023e8819-c436-4592-a6f6-ea3448733363","Type":"ContainerStarted","Data":"bd1108fbbb4023cda54606e07934f4d1186ed6393a32dfaca190fd006ca4009b"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.672769 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" event={"ID":"cfdab430-7dd4-4612-a65b-b381b5b58115","Type":"ContainerStarted","Data":"63f132cdc636e36f354d083d3f526984382664ee5b9a9bf902b35c34f2106e9d"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.681576 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" event={"ID":"e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9","Type":"ContainerStarted","Data":"a32aedefcc325dfc02325d178f024a500164b0dfeea7adffc0c11732128cf275"} Nov 23 20:21:47 crc kubenswrapper[4726]: E1123 20:21:47.686013 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" podUID="e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9" Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.714207 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" event={"ID":"4301d05b-4769-4c53-b557-bd6aa37efa68","Type":"ContainerStarted","Data":"e9d9afe537f31f9a1965be34085b75f67e8251c92437d41e70bfd3879d3c6155"} Nov 23 20:21:47 crc kubenswrapper[4726]: I1123 20:21:47.714251 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" event={"ID":"4301d05b-4769-4c53-b557-bd6aa37efa68","Type":"ContainerStarted","Data":"44f5960dcc549214610bd6603d4cc8f2e7d9750ef5159028ee8eba43dca058c8"} Nov 23 20:21:47 crc kubenswrapper[4726]: E1123 20:21:47.723660 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" podUID="4301d05b-4769-4c53-b557-bd6aa37efa68" Nov 23 20:21:48 crc kubenswrapper[4726]: I1123 20:21:48.780914 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" event={"ID":"706d33da-dac2-4403-a4d8-d8e7f04f772b","Type":"ContainerStarted","Data":"f744fef8d9423384551d35873f46bb559503703cfd65effcad9321d1972c813e"} Nov 23 20:21:48 crc kubenswrapper[4726]: I1123 20:21:48.780992 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" event={"ID":"706d33da-dac2-4403-a4d8-d8e7f04f772b","Type":"ContainerStarted","Data":"3b0bb0ee19e00c4f68a3eff75990731fc07b4a1593b728df18aca0c412279cbb"} Nov 23 20:21:48 crc kubenswrapper[4726]: I1123 20:21:48.781098 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:48 crc kubenswrapper[4726]: I1123 20:21:48.820045 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" event={"ID":"c22e9617-abf8-4370-ba10-31e967257270","Type":"ContainerStarted","Data":"bdf48a2c8338c566268ea13f37d0c256a4322f96aa22eaa30dd447ce8dfd7263"} Nov 23 20:21:48 crc kubenswrapper[4726]: E1123 20:21:48.825765 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" podUID="c22e9617-abf8-4370-ba10-31e967257270" Nov 23 20:21:48 crc kubenswrapper[4726]: E1123 20:21:48.826050 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" podUID="e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9" Nov 23 20:21:48 crc kubenswrapper[4726]: E1123 20:21:48.826095 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" podUID="4301d05b-4769-4c53-b557-bd6aa37efa68" Nov 23 20:21:48 crc kubenswrapper[4726]: I1123 20:21:48.843418 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" podStartSLOduration=4.843400781 podStartE2EDuration="4.843400781s" podCreationTimestamp="2025-11-23 20:21:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:21:48.83260514 +0000 UTC m=+816.981646096" watchObservedRunningTime="2025-11-23 20:21:48.843400781 +0000 UTC m=+816.992441727" Nov 23 20:21:49 crc kubenswrapper[4726]: I1123 20:21:49.024796 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:49 crc kubenswrapper[4726]: I1123 20:21:49.024841 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:49 crc kubenswrapper[4726]: I1123 20:21:49.167352 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:49 crc kubenswrapper[4726]: E1123 20:21:49.841063 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" podUID="c22e9617-abf8-4370-ba10-31e967257270" Nov 23 20:21:49 crc kubenswrapper[4726]: I1123 20:21:49.914916 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:49 crc kubenswrapper[4726]: I1123 20:21:49.996112 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kd7t5"] Nov 23 20:21:51 crc kubenswrapper[4726]: I1123 20:21:51.859226 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kd7t5" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" containerName="registry-server" containerID="cri-o://443f2466af374ec6ab66aba4901ffb04089b458155fc45edc59498224429efa5" gracePeriod=2 Nov 23 20:21:52 crc kubenswrapper[4726]: I1123 20:21:52.933291 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a8edd9c-fe94-4671-be59-31299be802fe" containerID="443f2466af374ec6ab66aba4901ffb04089b458155fc45edc59498224429efa5" exitCode=0 Nov 23 20:21:52 crc kubenswrapper[4726]: I1123 20:21:52.933373 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd7t5" event={"ID":"6a8edd9c-fe94-4671-be59-31299be802fe","Type":"ContainerDied","Data":"443f2466af374ec6ab66aba4901ffb04089b458155fc45edc59498224429efa5"} Nov 23 20:21:53 crc kubenswrapper[4726]: I1123 20:21:53.942977 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd7t5" event={"ID":"6a8edd9c-fe94-4671-be59-31299be802fe","Type":"ContainerDied","Data":"d72a628026d12caa36ff4e0639780236db26b5402a682961433cb76a132388d4"} Nov 23 20:21:53 crc kubenswrapper[4726]: I1123 20:21:53.943624 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d72a628026d12caa36ff4e0639780236db26b5402a682961433cb76a132388d4" Nov 23 20:21:53 crc kubenswrapper[4726]: I1123 20:21:53.978423 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.012524 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwpzr\" (UniqueName: \"kubernetes.io/projected/6a8edd9c-fe94-4671-be59-31299be802fe-kube-api-access-zwpzr\") pod \"6a8edd9c-fe94-4671-be59-31299be802fe\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.012587 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-utilities\") pod \"6a8edd9c-fe94-4671-be59-31299be802fe\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.012740 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-catalog-content\") pod \"6a8edd9c-fe94-4671-be59-31299be802fe\" (UID: \"6a8edd9c-fe94-4671-be59-31299be802fe\") " Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.013515 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-utilities" (OuterVolumeSpecName: "utilities") pod "6a8edd9c-fe94-4671-be59-31299be802fe" (UID: "6a8edd9c-fe94-4671-be59-31299be802fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.035783 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8edd9c-fe94-4671-be59-31299be802fe-kube-api-access-zwpzr" (OuterVolumeSpecName: "kube-api-access-zwpzr") pod "6a8edd9c-fe94-4671-be59-31299be802fe" (UID: "6a8edd9c-fe94-4671-be59-31299be802fe"). InnerVolumeSpecName "kube-api-access-zwpzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.079065 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a8edd9c-fe94-4671-be59-31299be802fe" (UID: "6a8edd9c-fe94-4671-be59-31299be802fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.114782 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.114819 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwpzr\" (UniqueName: \"kubernetes.io/projected/6a8edd9c-fe94-4671-be59-31299be802fe-kube-api-access-zwpzr\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.114837 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8edd9c-fe94-4671-be59-31299be802fe-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.948300 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd7t5" Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.971905 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kd7t5"] Nov 23 20:21:54 crc kubenswrapper[4726]: I1123 20:21:54.979462 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kd7t5"] Nov 23 20:21:56 crc kubenswrapper[4726]: I1123 20:21:56.015137 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-t9ptk" Nov 23 20:21:56 crc kubenswrapper[4726]: I1123 20:21:56.602653 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" path="/var/lib/kubelet/pods/6a8edd9c-fe94-4671-be59-31299be802fe/volumes" Nov 23 20:22:01 crc kubenswrapper[4726]: E1123 20:22:01.652060 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377" Nov 23 20:22:01 crc kubenswrapper[4726]: E1123 20:22:01.652704 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z64r2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-5c75d7c94b-m4xc4_openstack-operators(c53c8659-c283-4c78-88b4-b60f8b040dd7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:02 crc kubenswrapper[4726]: E1123 20:22:02.861287 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7" Nov 23 20:22:02 crc kubenswrapper[4726]: E1123 20:22:02.861536 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cqrsd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-86d796d84d-qvx6m_openstack-operators(384a1b22-97f2-4f0b-ad98-52adbbf5178b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:03 crc kubenswrapper[4726]: E1123 20:22:03.445189 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04" Nov 23 20:22:03 crc kubenswrapper[4726]: E1123 20:22:03.445367 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xdmbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6f8c5b86cb-8dz5d_openstack-operators(b89cb12d-b390-40ed-908f-e1095fe0c62b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:03 crc kubenswrapper[4726]: E1123 20:22:03.948930 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13" Nov 23 20:22:03 crc kubenswrapper[4726]: E1123 20:22:03.949482 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mvsdd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-6fdc856c5d-7lwwg_openstack-operators(20b0ca2b-a3ee-45c9-8b34-b476245a4ed6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:05 crc kubenswrapper[4726]: E1123 20:22:05.141133 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0" Nov 23 20:22:05 crc kubenswrapper[4726]: E1123 20:22:05.141360 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tnbp8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-799cb6ffd6-frss6_openstack-operators(cfdab430-7dd4-4612-a65b-b381b5b58115): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:05 crc kubenswrapper[4726]: E1123 20:22:05.628316 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96" Nov 23 20:22:05 crc kubenswrapper[4726]: E1123 20:22:05.628483 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cg2w4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-bf4c6585d-jc55l_openstack-operators(9c05c149-b530-4ca3-b201-309826c5e095): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:06 crc kubenswrapper[4726]: E1123 20:22:06.669320 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd" Nov 23 20:22:06 crc kubenswrapper[4726]: E1123 20:22:06.669804 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:7dbadf7b98f2f305f9f1382f55a084c8ca404f4263f76b28e56bd0dc437e2192,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:c8101c77a82eae4407e41e1fd766dfc6e1b7f9ed1679e3efb6f91ff97a1557b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:eb9743b21bbadca6f7cb9ac4fc46b5d58c51c674073c7e1121f4474a71304071,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:3d81f839b98c2e2a5bf0da79f2f9a92dff7d0a3c5a830b0e95c89dad8cf98a6a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:d19ac99249b47dd8ea16cd6aaa5756346aa8a2f119ee50819c15c5366efb417d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:8536169e5537fe6c330eba814248abdcf39cdd8f7e7336034d74e6fda9544050,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:4f1fa337760e82bfd67cdd142a97c121146dd7e621daac161940dd5e4ddb80dc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:3613b345d5baed98effd906f8b0242d863e14c97078ea473ef01fe1b0afc46f3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:9f9f367ed4c85efb16c3a74a4bb707ff0db271d7bc5abc70a71e984b55f43003,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:b73ad22b4955b06d584bce81742556d8c0c7828c495494f8ea7c99391c61b70f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:aa1d3aaf6b394621ed4089a98e0a82b763f467e8b5c5db772f9fdf99fc86e333,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:d6661053141b6df421288a7c9968a155ab82e478c1d75ab41f2cebe2f0ca02d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:ce2d63258cb4e7d0d1c07234de6889c5434464190906798019311a1c7cf6387f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:0485ef9e5b4437f7cd2ba54034a87722ce4669ee86b3773c6b0c037ed8000e91,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api@sha256:962c004551d0503779364b767b9bf0cecdf78dbba8809b2ca8b073f58e1f4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor@sha256:0ebf4c465fb6cc7dad9e6cb2da0ff54874c9acbcb40d62234a629ec2c12cdd62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:ff0c553ceeb2e0f44b010e37dc6d0db8a251797b88e56468b7cf7f05253e4232,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:624f553f073af7493d34828b074adc9981cce403edd8e71482c7307008479fd9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:e3874936a518c8560339db8f840fc5461885819f6050b5de8d3ab9199bea5094,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:1cea25f1d2a45affc80c46fb9d427749d3f06b61590ac6070a2910e3ec8a4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:e36d5b9a65194f12f7b01c6422ba3ed52a687fd1695fbb21f4986c67d9f9317f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:8b21bec527d54cd766e277889df6bcccd2baeaa946274606b986c0c3b7ca689f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:45aceca77f8fcf61127f0da650bdfdf11ede9b0944c78b63fab819d03283f96b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:709ac58998927dd61786821ae1e63343fd97ccf5763aac5edb4583eea9401d22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:867d4ef7c21f75e6030a685b5762ab4d84b671316ed6b98d75200076e93342cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2b90da93550b99d2fcfa95bd819f3363aa68346a416f8dc7baac3e9c5f487761,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:6f86db36d668348be8c5b46dcda8b1fa23d34bfdc07164fbcbe7a6327fb4de24,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:8cde52cef8795d1c91983b100d86541c7718160ec260fe0f97b96add4c2c8ee8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:a9583cb3baf440d2358ef041373833afbeae60da8159dd031502379901141620,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:835ebed082fe1c45bd799d1d5357595ce63efeb05ca876f26b08443facb9c164,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:011d682241db724bc40736c9b54d2ea450ea7e6be095b1ff5fa28c8007466775,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:2025da90cff8f563deb08bee71efe16d4078edc2a767b2e225cca5c77f1aa2f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:26bd7b0bd6070856aefef6fe754c547d55c056396ea30d879d34c2d49b5a1d29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:ff46cd5e0e13d105c4629e78c2734a50835f06b6a1e31da9e0462981d10c4be3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:5b4fd0c2b76fa5539f74687b11c5882d77bd31352452322b37ff51fa18f12a61,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:5e03376bd895346dc8f627ca15ded942526ed8b5e92872f453ce272e694d18d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:5f6045841aff0fde6f684a34cdf49f8dc7b2c3bcbdeab201f1058971e0c5f79e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:448f4e1b740c30936e340bd6e8534d78c83357bf373a4223950aa64d3484f007,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:b68e3615af8a0eb0ef6bf9ceeef59540a6f4a9a85f6078a3620be115c73a7db8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:7eae01cf60383e523c9cd94d158a9162120a7370829a1dad20fdea6b0fd660bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:28cc10501788081eb61b5a1af35546191a92741f4f109df54c74e2b19439d0f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:9a616e37acfd120612f78043237a8541266ba34883833c9beb43f3da313661ad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:6b1be6cd94a0942259bca5d5d2c30cc7de4a33276b61f8ae3940226772106256,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:02d2c22d15401574941fbe057095442dee0d6f7a0a9341de35d25e6a12a3fe4b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:fc3b3a36b74fd653946723c54b208072d52200635850b531e9d595a7aaea5a01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:7850ccbff320bf9a1c9c769c1c70777eb97117dd8cd5ae4435be9b4622cf807a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:397dac7e39cf40d14a986e6ec4a60fb698ca35c197d0db315b1318514cc6d1d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:1c95142a36276686e720f86423ee171dc9adcc1e89879f627545b7c906ccd9bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:e331a8fde6638e5ba154c4f0b38772a9a424f60656f2777245975fb1fa02f07d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:b6e1e8a249d36ef36c6ac4170af1e043dda1ccc0f9672832d3ff151bf3533076,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:cd3cf7a34053e850b4d4f9f4ea4c74953a54a42fd18e47d7c01d44a88923e925,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:aee28476344fc0cc148fbe97daf9b1bfcedc22001550bba4bdc4e84be7b6989d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:cfa0b92c976603ee2a937d34013a238fcd8aa75f998e50642e33489f14124633,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:73c2f2d6eecf88acf4e45b133c8373d9bb006b530e0aff0b28f3b7420620a874,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:927b405cc04abe5ff716186e8d35e2dc5fad1c8430194659ee6617d74e4e055d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:6154d7cebd7c339afa5b86330262156171743aa5b79c2b78f9a2f378005ed8fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:e2db2f4af8d3d0be7868c6efef0189f3a2c74a8f96ae10e3f991cdf83feaef29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:c773629df257726a6d3cacc24a6e4df0babcd7d37df04e6d14676a8da028b9c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:776211111e2e6493706dbc49a3ba44f31d1b947919313ed3a0f35810e304ec52,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:0a98e8f5c83522ca6c8e40c5e9561f6628d2d5e69f0e8a64279c541c989d3d8b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:7cccf24ad0a152f90ca39893064f48a1656950ee8142685a5d482c71f0bdc9f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:af46761060c7987e1dee5f14c06d85b46f12ad8e09c83d4246ab4e3a65dfda3e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:05450b48f6b5352b2686a26e933e8727748edae2ae9652d9164b7d7a1817c55a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:fc9c99eeef91523482bd8f92661b393287e1f2a24ad2ba9e33191f8de9af74cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:3e4ecc02b4b5e0860482a93599ba9ca598c5ce26c093c46e701f96fe51acb208,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:2346037e064861c7892690d2e8b3e1eea1a26ce3c3a11fda0b41301965bc828c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:c26c3ff9cabe3593ceb10006e782bf9391ac14785768ce9eec4f938c2d3cf228,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:daa45220bb1c47922d0917aa8fe423bb82b03a01429f1c9e37635e701e352d71,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:a80a074e227d3238bb6f285788a9e886ae7a5909ccbc5c19c93c369bdfe5b3b8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:58ac66ca1be01fe0157977bd79a26cde4d0de153edfaf4162367c924826b2ef4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:99a63770d80cc7c3afa1118b400972fb0e6bff5284a2eae781b12582ad79c29c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:9ee4d84529394afcd860f1a1186484560f02f08c15c37cac42a22473b7116d5f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:ea15fadda7b0439ec637edfaf6ea5dbf3e35fb3be012c7c5a31e722c90becb11,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qkjqv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-79d88dcd44hnx49_openstack-operators(ce7ed28d-d56b-44de-b207-7c47ab684921): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:07 crc kubenswrapper[4726]: E1123 20:22:07.807831 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9" Nov 23 20:22:07 crc kubenswrapper[4726]: E1123 20:22:07.808625 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m7smb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6d8fd67bf7-xt9t6_openstack-operators(a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:08 crc kubenswrapper[4726]: E1123 20:22:08.379581 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a" Nov 23 20:22:08 crc kubenswrapper[4726]: E1123 20:22:08.379747 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kgzbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7879fb76fd-s7746_openstack-operators(76e7ee34-d8e7-4d5b-b79e-acd11b2b0735): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:08 crc kubenswrapper[4726]: E1123 20:22:08.828791 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a" Nov 23 20:22:08 crc kubenswrapper[4726]: E1123 20:22:08.828974 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhgh2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7bb88cb858-5s9v2_openstack-operators(a653d1f5-fe23-4a55-9422-33964eeb8488): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:10 crc kubenswrapper[4726]: E1123 20:22:10.759916 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c" Nov 23 20:22:10 crc kubenswrapper[4726]: E1123 20:22:10.760343 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hl5xh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-6dc664666c-p7r99_openstack-operators(42d03b10-01f7-4963-a97b-f3d443a1cdf7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:11 crc kubenswrapper[4726]: E1123 20:22:11.886007 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894" Nov 23 20:22:11 crc kubenswrapper[4726]: E1123 20:22:11.886185 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vhvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-769d9c7585-w4lxk_openstack-operators(765e2946-d293-4840-baf2-96a478748a25): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:12 crc kubenswrapper[4726]: E1123 20:22:12.425411 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 23 20:22:12 crc kubenswrapper[4726]: E1123 20:22:12.425618 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b84np,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7cd4fb6f79-7x7b7_openstack-operators(023e8819-c436-4592-a6f6-ea3448733363): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:12 crc kubenswrapper[4726]: E1123 20:22:12.821696 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6" Nov 23 20:22:12 crc kubenswrapper[4726]: E1123 20:22:12.822107 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6npmp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-66b7d6f598-4p7kg_openstack-operators(4301d05b-4769-4c53-b557-bd6aa37efa68): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:22:12 crc kubenswrapper[4726]: E1123 20:22:12.823286 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" podUID="4301d05b-4769-4c53-b557-bd6aa37efa68" Nov 23 20:22:14 crc kubenswrapper[4726]: E1123 20:22:14.769899 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" podUID="9c05c149-b530-4ca3-b201-309826c5e095" Nov 23 20:22:14 crc kubenswrapper[4726]: E1123 20:22:14.868730 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" podUID="42d03b10-01f7-4963-a97b-f3d443a1cdf7" Nov 23 20:22:14 crc kubenswrapper[4726]: E1123 20:22:14.892838 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" podUID="76e7ee34-d8e7-4d5b-b79e-acd11b2b0735" Nov 23 20:22:14 crc kubenswrapper[4726]: E1123 20:22:14.896361 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" podUID="ce7ed28d-d56b-44de-b207-7c47ab684921" Nov 23 20:22:14 crc kubenswrapper[4726]: E1123 20:22:14.981600 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" podUID="b89cb12d-b390-40ed-908f-e1095fe0c62b" Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.107438 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" event={"ID":"74d488f0-9a23-4774-94e7-75605573ce43","Type":"ContainerStarted","Data":"9aabf3f0b1e05728e7de6fd63f9ed2c42fc3abb63d74ad94a50a647968c9161e"} Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.108414 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" event={"ID":"d481bfb6-f28c-49ed-8efe-7fc35dc6d608","Type":"ContainerStarted","Data":"fa1d44c6bd12256bf5ed69dd99febf58d259d1db16f257ed05f1443a166880e4"} Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.109255 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" event={"ID":"c3279a0f-e448-47b2-91b3-6de329d77cec","Type":"ContainerStarted","Data":"482ac411600b964e49506f7f8f3c8c349e1b22f3ae1c01ff7054b865911a2633"} Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.117663 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" event={"ID":"6e692743-eb35-4a91-97af-8d86cf694470","Type":"ContainerStarted","Data":"a65962543173ee813c66c451efd8a20cf404ff82c86c5130c4c40eb4cb9e5647"} Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.123608 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" event={"ID":"76e7ee34-d8e7-4d5b-b79e-acd11b2b0735","Type":"ContainerStarted","Data":"576e99950b72ad1d7a7e7d1f804937ca3f2f68e2a4f2a1f10fbc6a2237de00f9"} Nov 23 20:22:15 crc kubenswrapper[4726]: E1123 20:22:15.125108 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" podUID="76e7ee34-d8e7-4d5b-b79e-acd11b2b0735" Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.131097 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" event={"ID":"b89cb12d-b390-40ed-908f-e1095fe0c62b","Type":"ContainerStarted","Data":"30073d479a0d04d73776b6c7d5a0516a07e4205f968e4a25f43a59dc29815f61"} Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.134266 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" event={"ID":"bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb","Type":"ContainerStarted","Data":"f87aa8eb693c7697bb4e0ce257a4e800d4364190b29f5d9645f95f05ad0474db"} Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.144018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" event={"ID":"2e4ac3d0-65ba-4140-9a03-c269dfebad67","Type":"ContainerStarted","Data":"a6aa4798eec4c8682ab2bafadec2e14c14f7de96943daa2135d008eea6dd7fc0"} Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.147035 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" event={"ID":"9c05c149-b530-4ca3-b201-309826c5e095","Type":"ContainerStarted","Data":"a5218d88672c054121d186509044fdff2ab39b600336ea22fac2e85ab368ead9"} Nov 23 20:22:15 crc kubenswrapper[4726]: E1123 20:22:15.148409 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96\\\"\"" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" podUID="9c05c149-b530-4ca3-b201-309826c5e095" Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.154072 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" event={"ID":"ce7ed28d-d56b-44de-b207-7c47ab684921","Type":"ContainerStarted","Data":"75921e7c17b9fed92437dd6921b6d01d00205f01bcf7f3559662dc8b5c3f0e0a"} Nov 23 20:22:15 crc kubenswrapper[4726]: E1123 20:22:15.156672 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" podUID="ce7ed28d-d56b-44de-b207-7c47ab684921" Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.169679 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" event={"ID":"42d03b10-01f7-4963-a97b-f3d443a1cdf7","Type":"ContainerStarted","Data":"3635e18950c5bd70997c8ff632bcbcd4ddb38ea28f22d1b4460d1263791dfc8a"} Nov 23 20:22:15 crc kubenswrapper[4726]: E1123 20:22:15.170995 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" podUID="42d03b10-01f7-4963-a97b-f3d443a1cdf7" Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.183963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" event={"ID":"c22e9617-abf8-4370-ba10-31e967257270","Type":"ContainerStarted","Data":"a71c108499c01bb8173eb3685030577a0ac9a44d506600065cd005b50f7848f3"} Nov 23 20:22:15 crc kubenswrapper[4726]: I1123 20:22:15.184954 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.194168 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" podUID="76e7ee34-d8e7-4d5b-b79e-acd11b2b0735" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.194818 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" podUID="ce7ed28d-d56b-44de-b207-7c47ab684921" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.196301 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" podUID="42d03b10-01f7-4963-a97b-f3d443a1cdf7" Nov 23 20:22:16 crc kubenswrapper[4726]: I1123 20:22:16.221572 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" podStartSLOduration=5.133799878 podStartE2EDuration="32.221532627s" podCreationTimestamp="2025-11-23 20:21:44 +0000 UTC" firstStartedPulling="2025-11-23 20:21:47.157335724 +0000 UTC m=+815.306376680" lastFinishedPulling="2025-11-23 20:22:14.245068473 +0000 UTC m=+842.394109429" observedRunningTime="2025-11-23 20:22:15.3651315 +0000 UTC m=+843.514172466" watchObservedRunningTime="2025-11-23 20:22:16.221532627 +0000 UTC m=+844.370573583" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.478434 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" podUID="a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.479418 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" podUID="20b0ca2b-a3ee-45c9-8b34-b476245a4ed6" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.480212 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" podUID="c53c8659-c283-4c78-88b4-b60f8b040dd7" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.520275 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" podUID="a653d1f5-fe23-4a55-9422-33964eeb8488" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.520476 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" podUID="023e8819-c436-4592-a6f6-ea3448733363" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.520595 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" podUID="765e2946-d293-4840-baf2-96a478748a25" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.520683 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" podUID="384a1b22-97f2-4f0b-ad98-52adbbf5178b" Nov 23 20:22:16 crc kubenswrapper[4726]: E1123 20:22:16.543976 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" podUID="cfdab430-7dd4-4612-a65b-b381b5b58115" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.200493 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" event={"ID":"6e692743-eb35-4a91-97af-8d86cf694470","Type":"ContainerStarted","Data":"1aa3028d64fc0887f0644f8303a97b4dd8a15393deb1b37cde86cdc99f06aa4c"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.201531 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.202604 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" event={"ID":"384a1b22-97f2-4f0b-ad98-52adbbf5178b","Type":"ContainerStarted","Data":"d0fa58262b906cd680df40fd0879098b01637c9682dc37cf4aa7468b6fa0a815"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.205381 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" event={"ID":"a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7","Type":"ContainerStarted","Data":"2c972a2282a992ea1b194720aa76961f36eec74e855fca905bbb8fdbb838adfc"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.206671 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" event={"ID":"9c05c149-b530-4ca3-b201-309826c5e095","Type":"ContainerStarted","Data":"1b94c6a224ce93705fa84cbfd8f6bb88ce17b8378d7f1191b377067ab9993fe0"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.206898 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" Nov 23 20:22:17 crc kubenswrapper[4726]: E1123 20:22:17.207197 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" podUID="a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.208709 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" event={"ID":"d481bfb6-f28c-49ed-8efe-7fc35dc6d608","Type":"ContainerStarted","Data":"78db32f24ba17013a70158ccdd7f763dee21a6653bd0778bff567500f4642f42"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.208842 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.209860 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" event={"ID":"023e8819-c436-4592-a6f6-ea3448733363","Type":"ContainerStarted","Data":"b05b8a42a613e65015ba8f116bba11bfe67d14b2c13e1a5040aedc38bc333837"} Nov 23 20:22:17 crc kubenswrapper[4726]: E1123 20:22:17.211004 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" podUID="023e8819-c436-4592-a6f6-ea3448733363" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.211606 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" event={"ID":"e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9","Type":"ContainerStarted","Data":"b670f33cd8629e7e44ad03116a18df68310718d0efa62f8c5d63c73243700aec"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.213067 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" event={"ID":"20b0ca2b-a3ee-45c9-8b34-b476245a4ed6","Type":"ContainerStarted","Data":"467a30ce4a4f04cbf4c798767bf48e4cc0426a3767b4a81e36e3d8308d1d04e5"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.214912 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" event={"ID":"2e4ac3d0-65ba-4140-9a03-c269dfebad67","Type":"ContainerStarted","Data":"41a51177fb9c1b0cfafb32e8ee1b6ee1a3b02af342d0944f5b7997904aa0bb20"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.215106 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.218274 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" event={"ID":"765e2946-d293-4840-baf2-96a478748a25","Type":"ContainerStarted","Data":"93d548ed9ac40961f8d07faf754199ed7c296fabae7972c94f2cbde8c5e01034"} Nov 23 20:22:17 crc kubenswrapper[4726]: E1123 20:22:17.219361 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" podUID="765e2946-d293-4840-baf2-96a478748a25" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.222389 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" event={"ID":"b89cb12d-b390-40ed-908f-e1095fe0c62b","Type":"ContainerStarted","Data":"7e4c9b7e8458ac80a4e281bbca8a9e14021c374f6ebe322da96246cdad02a989"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.222690 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.224313 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" event={"ID":"a653d1f5-fe23-4a55-9422-33964eeb8488","Type":"ContainerStarted","Data":"302bbed2f87d675ff6f21a202e85081d8e71dafc4f4c89d0a6b21c86a63e99a4"} Nov 23 20:22:17 crc kubenswrapper[4726]: E1123 20:22:17.225330 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" podUID="a653d1f5-fe23-4a55-9422-33964eeb8488" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.227011 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" event={"ID":"c3279a0f-e448-47b2-91b3-6de329d77cec","Type":"ContainerStarted","Data":"c1cfd9c5f3946d02aa3faf24a28ff62ad8bb2f4e38756162ad655201dd0ae21a"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.227110 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.228589 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" event={"ID":"c53c8659-c283-4c78-88b4-b60f8b040dd7","Type":"ContainerStarted","Data":"84d8ef88681652acaf1136b07d36f107a7d629bc3d0cb1382be3d54c1ee3c846"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.230227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" event={"ID":"cfdab430-7dd4-4612-a65b-b381b5b58115","Type":"ContainerStarted","Data":"90fb115e5f8743d229d8d81733857160efa70e798d695f29b4c24a9de1269011"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.235233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" event={"ID":"74d488f0-9a23-4774-94e7-75605573ce43","Type":"ContainerStarted","Data":"519a5e47508c1005adc59b879eb306281a19fe2fd8ec248ce7016a002b56875c"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.235322 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.238148 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" event={"ID":"bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb","Type":"ContainerStarted","Data":"8e61159584204b82419578dc66caf2823c33f59195df2cd82f4e673fbe4b5d06"} Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.238330 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.252919 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" podStartSLOduration=5.9368126 podStartE2EDuration="33.252904781s" podCreationTimestamp="2025-11-23 20:21:44 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.907311059 +0000 UTC m=+815.056352005" lastFinishedPulling="2025-11-23 20:22:14.22340323 +0000 UTC m=+842.372444186" observedRunningTime="2025-11-23 20:22:17.252590972 +0000 UTC m=+845.401631928" watchObservedRunningTime="2025-11-23 20:22:17.252904781 +0000 UTC m=+845.401945737" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.277078 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" podStartSLOduration=3.714333395 podStartE2EDuration="34.277059997s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.115997884 +0000 UTC m=+814.265038840" lastFinishedPulling="2025-11-23 20:22:16.678724486 +0000 UTC m=+844.827765442" observedRunningTime="2025-11-23 20:22:17.276254283 +0000 UTC m=+845.425295239" watchObservedRunningTime="2025-11-23 20:22:17.277059997 +0000 UTC m=+845.426100953" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.342567 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" podStartSLOduration=6.052914002 podStartE2EDuration="33.342538061s" podCreationTimestamp="2025-11-23 20:21:44 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.936023395 +0000 UTC m=+815.085064351" lastFinishedPulling="2025-11-23 20:22:14.225647454 +0000 UTC m=+842.374688410" observedRunningTime="2025-11-23 20:22:17.342187701 +0000 UTC m=+845.491228647" watchObservedRunningTime="2025-11-23 20:22:17.342538061 +0000 UTC m=+845.491579017" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.385764 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" podStartSLOduration=4.6372390580000005 podStartE2EDuration="34.385746585s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.816128125 +0000 UTC m=+814.965169071" lastFinishedPulling="2025-11-23 20:22:16.564635642 +0000 UTC m=+844.713676598" observedRunningTime="2025-11-23 20:22:17.385100526 +0000 UTC m=+845.534141492" watchObservedRunningTime="2025-11-23 20:22:17.385746585 +0000 UTC m=+845.534787541" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.452617 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" podStartSLOduration=6.262778821 podStartE2EDuration="34.452593999s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:45.650075414 +0000 UTC m=+813.799116370" lastFinishedPulling="2025-11-23 20:22:13.839890552 +0000 UTC m=+841.988931548" observedRunningTime="2025-11-23 20:22:17.450522229 +0000 UTC m=+845.599563195" watchObservedRunningTime="2025-11-23 20:22:17.452593999 +0000 UTC m=+845.601634955" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.483254 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" podStartSLOduration=7.6475008639999995 podStartE2EDuration="34.48322758s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:45.969202899 +0000 UTC m=+814.118243855" lastFinishedPulling="2025-11-23 20:22:12.804929615 +0000 UTC m=+840.953970571" observedRunningTime="2025-11-23 20:22:17.476757164 +0000 UTC m=+845.625798130" watchObservedRunningTime="2025-11-23 20:22:17.48322758 +0000 UTC m=+845.632268566" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.518650 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv" podStartSLOduration=6.301192356 podStartE2EDuration="33.518628499s" podCreationTimestamp="2025-11-23 20:21:44 +0000 UTC" firstStartedPulling="2025-11-23 20:21:47.127051922 +0000 UTC m=+815.276092878" lastFinishedPulling="2025-11-23 20:22:14.344488065 +0000 UTC m=+842.493529021" observedRunningTime="2025-11-23 20:22:17.515656574 +0000 UTC m=+845.664697540" watchObservedRunningTime="2025-11-23 20:22:17.518628499 +0000 UTC m=+845.667669455" Nov 23 20:22:17 crc kubenswrapper[4726]: I1123 20:22:17.620489 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" podStartSLOduration=6.869584565 podStartE2EDuration="34.62046446s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.088926885 +0000 UTC m=+814.237967831" lastFinishedPulling="2025-11-23 20:22:13.83980677 +0000 UTC m=+841.988847726" observedRunningTime="2025-11-23 20:22:17.608027092 +0000 UTC m=+845.757068058" watchObservedRunningTime="2025-11-23 20:22:17.62046446 +0000 UTC m=+845.769505486" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.246120 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" event={"ID":"cfdab430-7dd4-4612-a65b-b381b5b58115","Type":"ContainerStarted","Data":"2a1ac30511a2d836bf6bca2df43d6f4a0ba1810fdd663a02a8215b7776bb70ea"} Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.246246 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.247551 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" event={"ID":"20b0ca2b-a3ee-45c9-8b34-b476245a4ed6","Type":"ContainerStarted","Data":"5442ed164642ee69b8819fda8715125fb52180c1d74b9cef6bc425d540075b82"} Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.247668 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.249597 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" event={"ID":"c53c8659-c283-4c78-88b4-b60f8b040dd7","Type":"ContainerStarted","Data":"3cd769bc6e8bc51593fcc63d632785d203ea2c8813c1362a36145fc0ddff9ccd"} Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.249711 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.251385 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" event={"ID":"384a1b22-97f2-4f0b-ad98-52adbbf5178b","Type":"ContainerStarted","Data":"5c81c665dcdc3c15c14d1378e59e2b1a1f13fab0c6827aa3286d80a0884ff7d6"} Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.252096 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" Nov 23 20:22:18 crc kubenswrapper[4726]: E1123 20:22:18.252633 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" podUID="765e2946-d293-4840-baf2-96a478748a25" Nov 23 20:22:18 crc kubenswrapper[4726]: E1123 20:22:18.252642 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" podUID="023e8819-c436-4592-a6f6-ea3448733363" Nov 23 20:22:18 crc kubenswrapper[4726]: E1123 20:22:18.254075 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" podUID="a653d1f5-fe23-4a55-9422-33964eeb8488" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.275812 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" podStartSLOduration=3.361892642 podStartE2EDuration="34.275790861s" podCreationTimestamp="2025-11-23 20:21:44 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.809534735 +0000 UTC m=+814.958575691" lastFinishedPulling="2025-11-23 20:22:17.723432954 +0000 UTC m=+845.872473910" observedRunningTime="2025-11-23 20:22:18.273088474 +0000 UTC m=+846.422129440" watchObservedRunningTime="2025-11-23 20:22:18.275790861 +0000 UTC m=+846.424831817" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.276349 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" podStartSLOduration=7.800387215 podStartE2EDuration="35.276344547s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:45.328977383 +0000 UTC m=+813.478018339" lastFinishedPulling="2025-11-23 20:22:12.804934715 +0000 UTC m=+840.953975671" observedRunningTime="2025-11-23 20:22:17.67117769 +0000 UTC m=+845.820218646" watchObservedRunningTime="2025-11-23 20:22:18.276344547 +0000 UTC m=+846.425385503" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.286546 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" podStartSLOduration=4.368802591 podStartE2EDuration="35.28652869s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.936452978 +0000 UTC m=+815.085493934" lastFinishedPulling="2025-11-23 20:22:17.854179087 +0000 UTC m=+846.003220033" observedRunningTime="2025-11-23 20:22:18.285562702 +0000 UTC m=+846.434603648" watchObservedRunningTime="2025-11-23 20:22:18.28652869 +0000 UTC m=+846.435569646" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.428424 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" podStartSLOduration=3.834860244 podStartE2EDuration="35.428406384s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.047966796 +0000 UTC m=+814.197007752" lastFinishedPulling="2025-11-23 20:22:17.641512936 +0000 UTC m=+845.790553892" observedRunningTime="2025-11-23 20:22:18.4234094 +0000 UTC m=+846.572450356" watchObservedRunningTime="2025-11-23 20:22:18.428406384 +0000 UTC m=+846.577447340" Nov 23 20:22:18 crc kubenswrapper[4726]: I1123 20:22:18.456223 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" podStartSLOduration=4.843173674 podStartE2EDuration="35.456207124s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.989322719 +0000 UTC m=+815.138363675" lastFinishedPulling="2025-11-23 20:22:17.602356169 +0000 UTC m=+845.751397125" observedRunningTime="2025-11-23 20:22:18.45363855 +0000 UTC m=+846.602679506" watchObservedRunningTime="2025-11-23 20:22:18.456207124 +0000 UTC m=+846.605248080" Nov 23 20:22:19 crc kubenswrapper[4726]: I1123 20:22:19.259916 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" event={"ID":"a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7","Type":"ContainerStarted","Data":"11fb5a41b5ff26b424e13b9bade897df666609d289ad0e0d752c84f6f49e1936"} Nov 23 20:22:19 crc kubenswrapper[4726]: I1123 20:22:19.260158 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" Nov 23 20:22:19 crc kubenswrapper[4726]: I1123 20:22:19.262238 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-2k5zx" Nov 23 20:22:19 crc kubenswrapper[4726]: I1123 20:22:19.277839 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" podStartSLOduration=2.941064779 podStartE2EDuration="36.277825261s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:45.324589866 +0000 UTC m=+813.473630822" lastFinishedPulling="2025-11-23 20:22:18.661350348 +0000 UTC m=+846.810391304" observedRunningTime="2025-11-23 20:22:19.274006551 +0000 UTC m=+847.423047527" watchObservedRunningTime="2025-11-23 20:22:19.277825261 +0000 UTC m=+847.426866217" Nov 23 20:22:23 crc kubenswrapper[4726]: E1123 20:22:23.591218 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" podUID="4301d05b-4769-4c53-b557-bd6aa37efa68" Nov 23 20:22:23 crc kubenswrapper[4726]: I1123 20:22:23.797072 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-q7qxp" Nov 23 20:22:23 crc kubenswrapper[4726]: I1123 20:22:23.802695 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-xt9t6" Nov 23 20:22:23 crc kubenswrapper[4726]: I1123 20:22:23.843885 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-wh58z" Nov 23 20:22:23 crc kubenswrapper[4726]: I1123 20:22:23.888517 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-6d8lj" Nov 23 20:22:23 crc kubenswrapper[4726]: I1123 20:22:23.932467 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-jc55l" Nov 23 20:22:24 crc kubenswrapper[4726]: I1123 20:22:24.011932 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-574zb" Nov 23 20:22:24 crc kubenswrapper[4726]: I1123 20:22:24.331474 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-m4xc4" Nov 23 20:22:24 crc kubenswrapper[4726]: I1123 20:22:24.417056 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-8dz5d" Nov 23 20:22:24 crc kubenswrapper[4726]: I1123 20:22:24.553407 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-qvx6m" Nov 23 20:22:24 crc kubenswrapper[4726]: I1123 20:22:24.693668 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-7lwwg" Nov 23 20:22:24 crc kubenswrapper[4726]: I1123 20:22:24.865031 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-frss6" Nov 23 20:22:24 crc kubenswrapper[4726]: I1123 20:22:24.888273 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-527bk" Nov 23 20:22:25 crc kubenswrapper[4726]: I1123 20:22:25.152215 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-8464cf66df-5w2nt" Nov 23 20:22:28 crc kubenswrapper[4726]: I1123 20:22:28.332317 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" event={"ID":"ce7ed28d-d56b-44de-b207-7c47ab684921","Type":"ContainerStarted","Data":"c56c27059f349828b660ec897e4e8555b08e4c20108fea2924e43be42f416e3a"} Nov 23 20:22:28 crc kubenswrapper[4726]: I1123 20:22:28.334214 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:22:28 crc kubenswrapper[4726]: I1123 20:22:28.380715 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" podStartSLOduration=4.487603809 podStartE2EDuration="45.380680399s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:47.197378916 +0000 UTC m=+815.346419872" lastFinishedPulling="2025-11-23 20:22:28.090455496 +0000 UTC m=+856.239496462" observedRunningTime="2025-11-23 20:22:28.370534137 +0000 UTC m=+856.519575103" watchObservedRunningTime="2025-11-23 20:22:28.380680399 +0000 UTC m=+856.529721355" Nov 23 20:22:29 crc kubenswrapper[4726]: I1123 20:22:29.343198 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" event={"ID":"76e7ee34-d8e7-4d5b-b79e-acd11b2b0735","Type":"ContainerStarted","Data":"870bac0bfbf300793dd685193b0e25b38b067b38d61bf9bba184de374a2e906a"} Nov 23 20:22:29 crc kubenswrapper[4726]: I1123 20:22:29.373119 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" podStartSLOduration=4.241788036 podStartE2EDuration="46.373094412s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.898512046 +0000 UTC m=+815.047553002" lastFinishedPulling="2025-11-23 20:22:29.029818382 +0000 UTC m=+857.178859378" observedRunningTime="2025-11-23 20:22:29.365193675 +0000 UTC m=+857.514234671" watchObservedRunningTime="2025-11-23 20:22:29.373094412 +0000 UTC m=+857.522135408" Nov 23 20:22:30 crc kubenswrapper[4726]: I1123 20:22:30.351838 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" event={"ID":"765e2946-d293-4840-baf2-96a478748a25","Type":"ContainerStarted","Data":"83fd00af3d6bd5b31abe987a27df44f03550842cb7c5a33d77401f87ab9ae2ad"} Nov 23 20:22:30 crc kubenswrapper[4726]: I1123 20:22:30.352132 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:22:30 crc kubenswrapper[4726]: I1123 20:22:30.372430 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" podStartSLOduration=4.089222264 podStartE2EDuration="47.372406222s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.818611936 +0000 UTC m=+814.967652892" lastFinishedPulling="2025-11-23 20:22:30.101795884 +0000 UTC m=+858.250836850" observedRunningTime="2025-11-23 20:22:30.366545414 +0000 UTC m=+858.515586420" watchObservedRunningTime="2025-11-23 20:22:30.372406222 +0000 UTC m=+858.521447188" Nov 23 20:22:31 crc kubenswrapper[4726]: I1123 20:22:31.360244 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" event={"ID":"023e8819-c436-4592-a6f6-ea3448733363","Type":"ContainerStarted","Data":"bbcdaedde1ceb7a8899c4701e305ed4f8078791087443909644ec7942874299b"} Nov 23 20:22:31 crc kubenswrapper[4726]: I1123 20:22:31.361399 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" Nov 23 20:22:31 crc kubenswrapper[4726]: I1123 20:22:31.376590 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" podStartSLOduration=3.149727415 podStartE2EDuration="47.376573724s" podCreationTimestamp="2025-11-23 20:21:44 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.780299583 +0000 UTC m=+814.929340539" lastFinishedPulling="2025-11-23 20:22:31.007145882 +0000 UTC m=+859.156186848" observedRunningTime="2025-11-23 20:22:31.373669801 +0000 UTC m=+859.522710767" watchObservedRunningTime="2025-11-23 20:22:31.376573724 +0000 UTC m=+859.525614680" Nov 23 20:22:32 crc kubenswrapper[4726]: I1123 20:22:32.368243 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" event={"ID":"a653d1f5-fe23-4a55-9422-33964eeb8488","Type":"ContainerStarted","Data":"9ca729faac74d1c074bfff9fe08119e7ef45b15e65d63f9050abb7896d486c97"} Nov 23 20:22:32 crc kubenswrapper[4726]: I1123 20:22:32.368437 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" Nov 23 20:22:32 crc kubenswrapper[4726]: I1123 20:22:32.370596 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" event={"ID":"42d03b10-01f7-4963-a97b-f3d443a1cdf7","Type":"ContainerStarted","Data":"e9ae0bfe7ecff98f79d9882ec688b3f38fdbd154797da399f48d28a3f9975a2e"} Nov 23 20:22:32 crc kubenswrapper[4726]: I1123 20:22:32.370798 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" Nov 23 20:22:32 crc kubenswrapper[4726]: I1123 20:22:32.388678 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" podStartSLOduration=4.2144518699999995 podStartE2EDuration="49.388650913s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.868036349 +0000 UTC m=+815.017077305" lastFinishedPulling="2025-11-23 20:22:32.042235392 +0000 UTC m=+860.191276348" observedRunningTime="2025-11-23 20:22:32.382449644 +0000 UTC m=+860.531490610" watchObservedRunningTime="2025-11-23 20:22:32.388650913 +0000 UTC m=+860.537691899" Nov 23 20:22:32 crc kubenswrapper[4726]: I1123 20:22:32.403003 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" podStartSLOduration=3.335970936 podStartE2EDuration="48.402983225s" podCreationTimestamp="2025-11-23 20:21:44 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.936306093 +0000 UTC m=+815.085347049" lastFinishedPulling="2025-11-23 20:22:32.003318392 +0000 UTC m=+860.152359338" observedRunningTime="2025-11-23 20:22:32.40037873 +0000 UTC m=+860.549419706" watchObservedRunningTime="2025-11-23 20:22:32.402983225 +0000 UTC m=+860.552024181" Nov 23 20:22:34 crc kubenswrapper[4726]: I1123 20:22:34.257714 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" Nov 23 20:22:34 crc kubenswrapper[4726]: I1123 20:22:34.260561 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-s7746" Nov 23 20:22:36 crc kubenswrapper[4726]: I1123 20:22:36.172743 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44hnx49" Nov 23 20:22:37 crc kubenswrapper[4726]: I1123 20:22:37.405679 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" event={"ID":"4301d05b-4769-4c53-b557-bd6aa37efa68","Type":"ContainerStarted","Data":"b20951d0701c0440680e242661e1e848872801ada121df181dde679c537eb420"} Nov 23 20:22:37 crc kubenswrapper[4726]: I1123 20:22:37.407017 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" Nov 23 20:22:39 crc kubenswrapper[4726]: I1123 20:22:39.043409 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:22:39 crc kubenswrapper[4726]: I1123 20:22:39.043972 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:22:44 crc kubenswrapper[4726]: I1123 20:22:44.280322 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-5s9v2" Nov 23 20:22:44 crc kubenswrapper[4726]: I1123 20:22:44.310244 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" podStartSLOduration=11.959403315 podStartE2EDuration="1m1.310218476s" podCreationTimestamp="2025-11-23 20:21:43 +0000 UTC" firstStartedPulling="2025-11-23 20:21:46.936804968 +0000 UTC m=+815.085845924" lastFinishedPulling="2025-11-23 20:22:36.287620119 +0000 UTC m=+864.436661085" observedRunningTime="2025-11-23 20:22:37.428443712 +0000 UTC m=+865.577484668" watchObservedRunningTime="2025-11-23 20:22:44.310218476 +0000 UTC m=+872.459259442" Nov 23 20:22:44 crc kubenswrapper[4726]: I1123 20:22:44.530425 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-4p7kg" Nov 23 20:22:44 crc kubenswrapper[4726]: I1123 20:22:44.675905 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-w4lxk" Nov 23 20:22:44 crc kubenswrapper[4726]: I1123 20:22:44.845177 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-p7r99" Nov 23 20:22:45 crc kubenswrapper[4726]: I1123 20:22:45.098606 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-7x7b7" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.103388 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-v9hsl"] Nov 23 20:23:00 crc kubenswrapper[4726]: E1123 20:23:00.106255 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" containerName="registry-server" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.106275 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" containerName="registry-server" Nov 23 20:23:00 crc kubenswrapper[4726]: E1123 20:23:00.106294 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" containerName="extract-content" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.106300 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" containerName="extract-content" Nov 23 20:23:00 crc kubenswrapper[4726]: E1123 20:23:00.106327 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" containerName="extract-utilities" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.106333 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" containerName="extract-utilities" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.106468 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8edd9c-fe94-4671-be59-31299be802fe" containerName="registry-server" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.107235 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.118567 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.118762 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8f69j" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.118934 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.119049 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.125264 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-v9hsl"] Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.160045 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28pmq\" (UniqueName: \"kubernetes.io/projected/42ac6696-5fc8-4278-8a76-33acaf607ab4-kube-api-access-28pmq\") pod \"dnsmasq-dns-7bdd77c89-v9hsl\" (UID: \"42ac6696-5fc8-4278-8a76-33acaf607ab4\") " pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.160106 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac6696-5fc8-4278-8a76-33acaf607ab4-config\") pod \"dnsmasq-dns-7bdd77c89-v9hsl\" (UID: \"42ac6696-5fc8-4278-8a76-33acaf607ab4\") " pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.190625 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6584b49599-ftqwl"] Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.202581 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-ftqwl"] Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.202915 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.208613 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.260616 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28pmq\" (UniqueName: \"kubernetes.io/projected/42ac6696-5fc8-4278-8a76-33acaf607ab4-kube-api-access-28pmq\") pod \"dnsmasq-dns-7bdd77c89-v9hsl\" (UID: \"42ac6696-5fc8-4278-8a76-33acaf607ab4\") " pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.260665 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac6696-5fc8-4278-8a76-33acaf607ab4-config\") pod \"dnsmasq-dns-7bdd77c89-v9hsl\" (UID: \"42ac6696-5fc8-4278-8a76-33acaf607ab4\") " pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.260727 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-dns-svc\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.260743 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-config\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.260778 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n64h\" (UniqueName: \"kubernetes.io/projected/12c8453d-da78-4c44-a770-f1546072da33-kube-api-access-9n64h\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.262024 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac6696-5fc8-4278-8a76-33acaf607ab4-config\") pod \"dnsmasq-dns-7bdd77c89-v9hsl\" (UID: \"42ac6696-5fc8-4278-8a76-33acaf607ab4\") " pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.285978 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28pmq\" (UniqueName: \"kubernetes.io/projected/42ac6696-5fc8-4278-8a76-33acaf607ab4-kube-api-access-28pmq\") pod \"dnsmasq-dns-7bdd77c89-v9hsl\" (UID: \"42ac6696-5fc8-4278-8a76-33acaf607ab4\") " pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.361657 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-dns-svc\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.362259 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-config\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.362421 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n64h\" (UniqueName: \"kubernetes.io/projected/12c8453d-da78-4c44-a770-f1546072da33-kube-api-access-9n64h\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.362563 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-dns-svc\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.363057 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-config\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.393697 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n64h\" (UniqueName: \"kubernetes.io/projected/12c8453d-da78-4c44-a770-f1546072da33-kube-api-access-9n64h\") pod \"dnsmasq-dns-6584b49599-ftqwl\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.432720 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.517881 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.741788 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-v9hsl"] Nov 23 20:23:00 crc kubenswrapper[4726]: I1123 20:23:00.781280 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 20:23:01 crc kubenswrapper[4726]: I1123 20:23:01.147940 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-ftqwl"] Nov 23 20:23:01 crc kubenswrapper[4726]: W1123 20:23:01.150917 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12c8453d_da78_4c44_a770_f1546072da33.slice/crio-f079fb412da40dd2ed1e3ba38132983fb19d685407337d3a49be1ab046e1b796 WatchSource:0}: Error finding container f079fb412da40dd2ed1e3ba38132983fb19d685407337d3a49be1ab046e1b796: Status 404 returned error can't find the container with id f079fb412da40dd2ed1e3ba38132983fb19d685407337d3a49be1ab046e1b796 Nov 23 20:23:01 crc kubenswrapper[4726]: I1123 20:23:01.618672 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" event={"ID":"42ac6696-5fc8-4278-8a76-33acaf607ab4","Type":"ContainerStarted","Data":"bb640a1fb90b73161195d8f6143f4ab096b5c254dcc108bae74fa70cc85c6e73"} Nov 23 20:23:01 crc kubenswrapper[4726]: I1123 20:23:01.620819 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-ftqwl" event={"ID":"12c8453d-da78-4c44-a770-f1546072da33","Type":"ContainerStarted","Data":"f079fb412da40dd2ed1e3ba38132983fb19d685407337d3a49be1ab046e1b796"} Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.178086 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-ftqwl"] Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.208004 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-rq5b9"] Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.209057 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.292594 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-rq5b9"] Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.310383 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.310455 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr996\" (UniqueName: \"kubernetes.io/projected/31640635-e6ea-4445-bf46-007d16c78378-kube-api-access-jr996\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.313157 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-config\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.417571 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.417640 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr996\" (UniqueName: \"kubernetes.io/projected/31640635-e6ea-4445-bf46-007d16c78378-kube-api-access-jr996\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.417687 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-config\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.418818 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.425526 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-config\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.462973 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr996\" (UniqueName: \"kubernetes.io/projected/31640635-e6ea-4445-bf46-007d16c78378-kube-api-access-jr996\") pod \"dnsmasq-dns-7c6d9948dc-rq5b9\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.546349 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.707634 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-v9hsl"] Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.754034 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-khj6k"] Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.755326 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.784538 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-khj6k"] Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.930256 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-dns-svc\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.930316 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-config\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:03 crc kubenswrapper[4726]: I1123 20:23:03.930349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwt6z\" (UniqueName: \"kubernetes.io/projected/d2275f49-16fd-463d-9d90-314581001af4-kube-api-access-wwt6z\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.031582 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-dns-svc\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.031647 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-config\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.031681 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwt6z\" (UniqueName: \"kubernetes.io/projected/d2275f49-16fd-463d-9d90-314581001af4-kube-api-access-wwt6z\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.032948 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-dns-svc\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.033046 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-config\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.073303 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwt6z\" (UniqueName: \"kubernetes.io/projected/d2275f49-16fd-463d-9d90-314581001af4-kube-api-access-wwt6z\") pod \"dnsmasq-dns-6486446b9f-khj6k\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.087861 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.350773 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-rq5b9"] Nov 23 20:23:04 crc kubenswrapper[4726]: W1123 20:23:04.377101 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31640635_e6ea_4445_bf46_007d16c78378.slice/crio-0b26a12bfb1b211628f6d29b8d1f66e70e21f2ff436f7e45e2cd0ab1c075b997 WatchSource:0}: Error finding container 0b26a12bfb1b211628f6d29b8d1f66e70e21f2ff436f7e45e2cd0ab1c075b997: Status 404 returned error can't find the container with id 0b26a12bfb1b211628f6d29b8d1f66e70e21f2ff436f7e45e2cd0ab1c075b997 Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.415478 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.422197 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.431479 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.431760 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.431953 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.431995 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.432078 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.432116 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.432151 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-kmlks" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.434300 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543798 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543817 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-config-data\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543840 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543886 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1879312-c440-415f-a376-4ab58b99e21d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543914 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543940 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543962 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.543983 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.544020 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1879312-c440-415f-a376-4ab58b99e21d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.544040 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x4ht\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-kube-api-access-7x4ht\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.636027 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-khj6k"] Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.645359 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646159 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646195 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646221 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646315 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1879312-c440-415f-a376-4ab58b99e21d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646338 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x4ht\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-kube-api-access-7x4ht\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646373 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646397 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646424 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-config-data\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646449 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.646480 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1879312-c440-415f-a376-4ab58b99e21d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.647565 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-config-data\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.647966 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.648172 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.648826 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.649460 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.650173 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.657901 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1879312-c440-415f-a376-4ab58b99e21d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.666620 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.670329 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.671498 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1879312-c440-415f-a376-4ab58b99e21d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.680592 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x4ht\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-kube-api-access-7x4ht\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.724810 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" event={"ID":"31640635-e6ea-4445-bf46-007d16c78378","Type":"ContainerStarted","Data":"0b26a12bfb1b211628f6d29b8d1f66e70e21f2ff436f7e45e2cd0ab1c075b997"} Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.725062 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.813582 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.912723 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.916486 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.918013 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-24tl5" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.919021 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.921172 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.921207 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.921848 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.922632 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.922761 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 23 20:23:04 crc kubenswrapper[4726]: I1123 20:23:04.954554 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.059837 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.059927 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.059949 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.059979 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjwjq\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-kube-api-access-hjwjq\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.060001 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.060032 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.060056 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.060076 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.060102 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.060165 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.060197 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162229 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162619 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162649 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjwjq\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-kube-api-access-hjwjq\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162668 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162696 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162713 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162733 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162753 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162810 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162833 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.162861 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.164498 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.164716 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.165373 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.165674 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.169006 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.169303 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.171779 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.180653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.180771 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.188411 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.191195 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjwjq\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-kube-api-access-hjwjq\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.201717 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.268538 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.521613 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.769451 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1879312-c440-415f-a376-4ab58b99e21d","Type":"ContainerStarted","Data":"5027f5618fa01a06cafc6319af10587ee946bc63fb08f7432605af486edad51c"} Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.771222 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-khj6k" event={"ID":"d2275f49-16fd-463d-9d90-314581001af4","Type":"ContainerStarted","Data":"375157895b3950b8b64a35fbd4aa3b296f69b8babfcec8c8c00bb9c5b86226f9"} Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.844169 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.849363 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.856567 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.857672 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.858370 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.860064 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-g8cf2" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.862128 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.864057 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.982259 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-kolla-config\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.982308 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.982363 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.982381 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.982398 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g9vh\" (UniqueName: \"kubernetes.io/projected/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-kube-api-access-2g9vh\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.982417 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-config-data-default\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.982436 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:05 crc kubenswrapper[4726]: I1123 20:23:05.982457 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.025016 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.083419 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.083461 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.083521 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-kolla-config\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.083554 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.083617 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.083637 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.083651 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g9vh\" (UniqueName: \"kubernetes.io/projected/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-kube-api-access-2g9vh\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.083672 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-config-data-default\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.085652 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.086584 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-config-data-default\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.086750 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.089088 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.089472 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-kolla-config\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.094213 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.111548 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.126881 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g9vh\" (UniqueName: \"kubernetes.io/projected/3647d686-8ad6-4bad-9ca6-ffbaba9bba97-kube-api-access-2g9vh\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.155159 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"3647d686-8ad6-4bad-9ca6-ffbaba9bba97\") " pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.192234 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 23 20:23:06 crc kubenswrapper[4726]: I1123 20:23:06.812989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6","Type":"ContainerStarted","Data":"c2f9f0dd41e5163141b323218ea9ec9de39e4bed74ef0d6efd23b88f65bbf864"} Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.560033 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.560954 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.577886 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.578066 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.578172 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-knd5b" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.611061 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.614184 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.624097 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.624634 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.624755 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-2xwmm" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.625122 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.634821 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.640504 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727003 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727056 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-config-data\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727113 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727132 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gsl7\" (UniqueName: \"kubernetes.io/projected/e88e0fb1-897d-454c-9a23-8187592296e5-kube-api-access-9gsl7\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727167 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e88e0fb1-897d-454c-9a23-8187592296e5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57vxx\" (UniqueName: \"kubernetes.io/projected/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-kube-api-access-57vxx\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727211 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-kolla-config\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727230 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727260 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727298 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727319 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e88e0fb1-897d-454c-9a23-8187592296e5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727338 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e88e0fb1-897d-454c-9a23-8187592296e5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.727377 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828678 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e88e0fb1-897d-454c-9a23-8187592296e5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828720 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57vxx\" (UniqueName: \"kubernetes.io/projected/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-kube-api-access-57vxx\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828747 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-kolla-config\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828770 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828796 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828829 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828850 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e88e0fb1-897d-454c-9a23-8187592296e5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828894 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e88e0fb1-897d-454c-9a23-8187592296e5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828931 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828958 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.828985 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-config-data\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.829035 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.829051 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gsl7\" (UniqueName: \"kubernetes.io/projected/e88e0fb1-897d-454c-9a23-8187592296e5-kube-api-access-9gsl7\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.832533 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.834372 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e88e0fb1-897d-454c-9a23-8187592296e5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.835051 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.835782 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-kolla-config\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.837588 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-config-data\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.837939 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.838640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.838452 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e88e0fb1-897d-454c-9a23-8187592296e5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.839229 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e88e0fb1-897d-454c-9a23-8187592296e5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.851074 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e88e0fb1-897d-454c-9a23-8187592296e5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.854310 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.862377 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gsl7\" (UniqueName: \"kubernetes.io/projected/e88e0fb1-897d-454c-9a23-8187592296e5-kube-api-access-9gsl7\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.894660 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57vxx\" (UniqueName: \"kubernetes.io/projected/319fe004-3d9c-4bb9-bb8e-114c8949bcc7-kube-api-access-57vxx\") pod \"memcached-0\" (UID: \"319fe004-3d9c-4bb9-bb8e-114c8949bcc7\") " pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.905219 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.921173 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e88e0fb1-897d-454c-9a23-8187592296e5\") " pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:07 crc kubenswrapper[4726]: I1123 20:23:07.954274 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:08 crc kubenswrapper[4726]: I1123 20:23:08.068748 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 23 20:23:08 crc kubenswrapper[4726]: I1123 20:23:08.516210 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 23 20:23:08 crc kubenswrapper[4726]: W1123 20:23:08.551538 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod319fe004_3d9c_4bb9_bb8e_114c8949bcc7.slice/crio-1490ea52eecf33df9fd22ad49aea0cef40a517e68f1d4255b41b233c8142b81b WatchSource:0}: Error finding container 1490ea52eecf33df9fd22ad49aea0cef40a517e68f1d4255b41b233c8142b81b: Status 404 returned error can't find the container with id 1490ea52eecf33df9fd22ad49aea0cef40a517e68f1d4255b41b233c8142b81b Nov 23 20:23:08 crc kubenswrapper[4726]: I1123 20:23:08.663482 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 23 20:23:08 crc kubenswrapper[4726]: W1123 20:23:08.673185 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode88e0fb1_897d_454c_9a23_8187592296e5.slice/crio-326bf6705315de2f0edcb1000b60f34c332cb4ede68f01abe340f06804e89ed8 WatchSource:0}: Error finding container 326bf6705315de2f0edcb1000b60f34c332cb4ede68f01abe340f06804e89ed8: Status 404 returned error can't find the container with id 326bf6705315de2f0edcb1000b60f34c332cb4ede68f01abe340f06804e89ed8 Nov 23 20:23:08 crc kubenswrapper[4726]: I1123 20:23:08.940186 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3647d686-8ad6-4bad-9ca6-ffbaba9bba97","Type":"ContainerStarted","Data":"b12c83971e0f70b5f756da5a343f64937ba69cdf336b56a0ff59722f2e51738f"} Nov 23 20:23:08 crc kubenswrapper[4726]: I1123 20:23:08.943777 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e88e0fb1-897d-454c-9a23-8187592296e5","Type":"ContainerStarted","Data":"326bf6705315de2f0edcb1000b60f34c332cb4ede68f01abe340f06804e89ed8"} Nov 23 20:23:08 crc kubenswrapper[4726]: I1123 20:23:08.961920 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"319fe004-3d9c-4bb9-bb8e-114c8949bcc7","Type":"ContainerStarted","Data":"1490ea52eecf33df9fd22ad49aea0cef40a517e68f1d4255b41b233c8142b81b"} Nov 23 20:23:09 crc kubenswrapper[4726]: I1123 20:23:09.047051 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:23:09 crc kubenswrapper[4726]: I1123 20:23:09.047118 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:23:09 crc kubenswrapper[4726]: I1123 20:23:09.743654 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:23:09 crc kubenswrapper[4726]: I1123 20:23:09.744645 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 20:23:09 crc kubenswrapper[4726]: I1123 20:23:09.747232 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4hdwq" Nov 23 20:23:09 crc kubenswrapper[4726]: I1123 20:23:09.769626 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:23:09 crc kubenswrapper[4726]: I1123 20:23:09.924767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbb2j\" (UniqueName: \"kubernetes.io/projected/91c39827-8843-4434-8e7f-1ff5be6d9c24-kube-api-access-fbb2j\") pod \"kube-state-metrics-0\" (UID: \"91c39827-8843-4434-8e7f-1ff5be6d9c24\") " pod="openstack/kube-state-metrics-0" Nov 23 20:23:10 crc kubenswrapper[4726]: I1123 20:23:10.026928 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbb2j\" (UniqueName: \"kubernetes.io/projected/91c39827-8843-4434-8e7f-1ff5be6d9c24-kube-api-access-fbb2j\") pod \"kube-state-metrics-0\" (UID: \"91c39827-8843-4434-8e7f-1ff5be6d9c24\") " pod="openstack/kube-state-metrics-0" Nov 23 20:23:10 crc kubenswrapper[4726]: I1123 20:23:10.066425 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbb2j\" (UniqueName: \"kubernetes.io/projected/91c39827-8843-4434-8e7f-1ff5be6d9c24-kube-api-access-fbb2j\") pod \"kube-state-metrics-0\" (UID: \"91c39827-8843-4434-8e7f-1ff5be6d9c24\") " pod="openstack/kube-state-metrics-0" Nov 23 20:23:10 crc kubenswrapper[4726]: I1123 20:23:10.085024 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 20:23:11 crc kubenswrapper[4726]: I1123 20:23:11.070930 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:23:12 crc kubenswrapper[4726]: I1123 20:23:12.029930 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"91c39827-8843-4434-8e7f-1ff5be6d9c24","Type":"ContainerStarted","Data":"a80199e7223b40bfb77f2ae14fce97328f16719b49cd6b3c78a04c2786f6b8b6"} Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.555350 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-657sc"] Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.556714 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.589840 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-qzjmb" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.590215 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.597965 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.598506 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-657sc"] Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.606570 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-sq7pg"] Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.608551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.651980 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-run\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.652032 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-run\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.652052 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cb17e51-7e2f-4012-96fd-5f6bd424429c-scripts\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.652090 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cb17e51-7e2f-4012-96fd-5f6bd424429c-combined-ca-bundle\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.652111 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-etc-ovs\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.652185 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-log\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.652213 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-run-ovn\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.653514 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-lib\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.653563 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac53da75-5263-4d97-aa49-b4e8b345c854-scripts\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.653596 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzc72\" (UniqueName: \"kubernetes.io/projected/4cb17e51-7e2f-4012-96fd-5f6bd424429c-kube-api-access-lzc72\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.653618 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cb17e51-7e2f-4012-96fd-5f6bd424429c-ovn-controller-tls-certs\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.653651 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wh6f\" (UniqueName: \"kubernetes.io/projected/ac53da75-5263-4d97-aa49-b4e8b345c854-kube-api-access-9wh6f\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.653666 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-log-ovn\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.658887 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sq7pg"] Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755210 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-log\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755260 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-run-ovn\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755299 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-lib\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755327 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac53da75-5263-4d97-aa49-b4e8b345c854-scripts\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755351 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzc72\" (UniqueName: \"kubernetes.io/projected/4cb17e51-7e2f-4012-96fd-5f6bd424429c-kube-api-access-lzc72\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755442 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cb17e51-7e2f-4012-96fd-5f6bd424429c-ovn-controller-tls-certs\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755480 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wh6f\" (UniqueName: \"kubernetes.io/projected/ac53da75-5263-4d97-aa49-b4e8b345c854-kube-api-access-9wh6f\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755499 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-log-ovn\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.755547 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-run\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.756159 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-run\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.756187 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cb17e51-7e2f-4012-96fd-5f6bd424429c-scripts\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.756234 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cb17e51-7e2f-4012-96fd-5f6bd424429c-combined-ca-bundle\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.756258 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-etc-ovs\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.756754 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-etc-ovs\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.756916 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-log\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.757029 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-run-ovn\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.757123 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-lib\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.759003 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-log-ovn\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.759256 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4cb17e51-7e2f-4012-96fd-5f6bd424429c-var-run\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.759337 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac53da75-5263-4d97-aa49-b4e8b345c854-var-run\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.759576 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac53da75-5263-4d97-aa49-b4e8b345c854-scripts\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.761847 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4cb17e51-7e2f-4012-96fd-5f6bd424429c-scripts\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.772212 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cb17e51-7e2f-4012-96fd-5f6bd424429c-ovn-controller-tls-certs\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.774925 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cb17e51-7e2f-4012-96fd-5f6bd424429c-combined-ca-bundle\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.796809 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzc72\" (UniqueName: \"kubernetes.io/projected/4cb17e51-7e2f-4012-96fd-5f6bd424429c-kube-api-access-lzc72\") pod \"ovn-controller-657sc\" (UID: \"4cb17e51-7e2f-4012-96fd-5f6bd424429c\") " pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.810546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wh6f\" (UniqueName: \"kubernetes.io/projected/ac53da75-5263-4d97-aa49-b4e8b345c854-kube-api-access-9wh6f\") pod \"ovn-controller-ovs-sq7pg\" (UID: \"ac53da75-5263-4d97-aa49-b4e8b345c854\") " pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.909296 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc" Nov 23 20:23:13 crc kubenswrapper[4726]: I1123 20:23:13.945612 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.283819 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.291653 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.297666 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.309149 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.310275 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.311649 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tpr79" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.315911 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.316324 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.394162 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.394255 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.394298 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h97jf\" (UniqueName: \"kubernetes.io/projected/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-kube-api-access-h97jf\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.394327 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-config\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.394346 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.394391 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.394411 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.394428 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.497837 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.497956 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.497982 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.498502 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.498906 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.499436 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.499465 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.499612 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.499720 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h97jf\" (UniqueName: \"kubernetes.io/projected/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-kube-api-access-h97jf\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.499749 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-config\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.499772 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.500612 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-config\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.518700 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.519505 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.520360 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.523494 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h97jf\" (UniqueName: \"kubernetes.io/projected/ff67ad55-4c63-45f3-b6f5-12f941bdd2a7-kube-api-access-h97jf\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.527494 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7\") " pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:14 crc kubenswrapper[4726]: I1123 20:23:14.613630 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.089608 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.102855 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.103124 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.125754 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.125974 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.126094 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.126248 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-v8gl9" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.256835 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.256988 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.257050 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1cb13d69-d5a0-4a7b-929b-ed077760725a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.257106 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px5d9\" (UniqueName: \"kubernetes.io/projected/1cb13d69-d5a0-4a7b-929b-ed077760725a-kube-api-access-px5d9\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.257136 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.257152 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cb13d69-d5a0-4a7b-929b-ed077760725a-config\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.257173 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cb13d69-d5a0-4a7b-929b-ed077760725a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.257197 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358221 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358266 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cb13d69-d5a0-4a7b-929b-ed077760725a-config\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358291 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cb13d69-d5a0-4a7b-929b-ed077760725a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358322 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358353 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358371 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358420 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1cb13d69-d5a0-4a7b-929b-ed077760725a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358474 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px5d9\" (UniqueName: \"kubernetes.io/projected/1cb13d69-d5a0-4a7b-929b-ed077760725a-kube-api-access-px5d9\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.358577 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.359958 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cb13d69-d5a0-4a7b-929b-ed077760725a-config\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.359967 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cb13d69-d5a0-4a7b-929b-ed077760725a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.360400 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1cb13d69-d5a0-4a7b-929b-ed077760725a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.367198 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.368958 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.371584 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb13d69-d5a0-4a7b-929b-ed077760725a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.380364 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.383811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px5d9\" (UniqueName: \"kubernetes.io/projected/1cb13d69-d5a0-4a7b-929b-ed077760725a-kube-api-access-px5d9\") pod \"ovsdbserver-sb-0\" (UID: \"1cb13d69-d5a0-4a7b-929b-ed077760725a\") " pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:17 crc kubenswrapper[4726]: I1123 20:23:17.449914 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:19 crc kubenswrapper[4726]: I1123 20:23:19.527892 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-657sc"] Nov 23 20:23:20 crc kubenswrapper[4726]: I1123 20:23:20.065783 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 23 20:23:26 crc kubenswrapper[4726]: E1123 20:23:26.840034 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2" Nov 23 20:23:26 crc kubenswrapper[4726]: E1123 20:23:26.841103 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n544h76h59bh6chfch5cdh54ch5bdh64dh594h5d6h55dh58h644h7fh6ch84h685h67fh8bh699h99h687h647hbh5bhb4h64bh579h55bh54bh94q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-57vxx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(319fe004-3d9c-4bb9-bb8e-114c8949bcc7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:23:26 crc kubenswrapper[4726]: E1123 20:23:26.842367 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="319fe004-3d9c-4bb9-bb8e-114c8949bcc7" Nov 23 20:23:27 crc kubenswrapper[4726]: E1123 20:23:27.287794 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2\\\"\"" pod="openstack/memcached-0" podUID="319fe004-3d9c-4bb9-bb8e-114c8949bcc7" Nov 23 20:23:30 crc kubenswrapper[4726]: W1123 20:23:30.831257 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cb17e51_7e2f_4012_96fd_5f6bd424429c.slice/crio-d1680890a683e723cc9c61c63e2b381b1e12bcb6c75944cbcd42a82aeb3a682e WatchSource:0}: Error finding container d1680890a683e723cc9c61c63e2b381b1e12bcb6c75944cbcd42a82aeb3a682e: Status 404 returned error can't find the container with id d1680890a683e723cc9c61c63e2b381b1e12bcb6c75944cbcd42a82aeb3a682e Nov 23 20:23:30 crc kubenswrapper[4726]: W1123 20:23:30.845197 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff67ad55_4c63_45f3_b6f5_12f941bdd2a7.slice/crio-9c6c3c3d822bf771e40eaba6bc7a822e07ee29164c3b7736bb41e60c521ab89b WatchSource:0}: Error finding container 9c6c3c3d822bf771e40eaba6bc7a822e07ee29164c3b7736bb41e60c521ab89b: Status 404 returned error can't find the container with id 9c6c3c3d822bf771e40eaba6bc7a822e07ee29164c3b7736bb41e60c521ab89b Nov 23 20:23:31 crc kubenswrapper[4726]: I1123 20:23:31.333192 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7","Type":"ContainerStarted","Data":"9c6c3c3d822bf771e40eaba6bc7a822e07ee29164c3b7736bb41e60c521ab89b"} Nov 23 20:23:31 crc kubenswrapper[4726]: I1123 20:23:31.336001 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-657sc" event={"ID":"4cb17e51-7e2f-4012-96fd-5f6bd424429c","Type":"ContainerStarted","Data":"d1680890a683e723cc9c61c63e2b381b1e12bcb6c75944cbcd42a82aeb3a682e"} Nov 23 20:23:31 crc kubenswrapper[4726]: I1123 20:23:31.405150 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sq7pg"] Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.559053 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.559853 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jr996,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7c6d9948dc-rq5b9_openstack(31640635-e6ea-4445-bf46-007d16c78378): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.561046 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" podUID="31640635-e6ea-4445-bf46-007d16c78378" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.624174 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.624365 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wwt6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6486446b9f-khj6k_openstack(d2275f49-16fd-463d-9d90-314581001af4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.625440 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6486446b9f-khj6k" podUID="d2275f49-16fd-463d-9d90-314581001af4" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.635041 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.635232 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9n64h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6584b49599-ftqwl_openstack(12c8453d-da78-4c44-a770-f1546072da33): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.636332 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6584b49599-ftqwl" podUID="12c8453d-da78-4c44-a770-f1546072da33" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.649262 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.649775 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-28pmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bdd77c89-v9hsl_openstack(42ac6696-5fc8-4278-8a76-33acaf607ab4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:23:35 crc kubenswrapper[4726]: E1123 20:23:35.651102 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" podUID="42ac6696-5fc8-4278-8a76-33acaf607ab4" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.128377 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.370168 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sq7pg" event={"ID":"ac53da75-5263-4d97-aa49-b4e8b345c854","Type":"ContainerStarted","Data":"9181e1cf1524cf6abb276e33ae2548740e91c5a951049fc3bc1f46f2f780a7a9"} Nov 23 20:23:36 crc kubenswrapper[4726]: E1123 20:23:36.372955 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" podUID="31640635-e6ea-4445-bf46-007d16c78378" Nov 23 20:23:36 crc kubenswrapper[4726]: E1123 20:23:36.373069 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-6486446b9f-khj6k" podUID="d2275f49-16fd-463d-9d90-314581001af4" Nov 23 20:23:36 crc kubenswrapper[4726]: W1123 20:23:36.471180 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cb13d69_d5a0_4a7b_929b_ed077760725a.slice/crio-207f3492f26c495cef6bbffa4030a15f9f871fd2a98f20dbcf685d7a52eceb56 WatchSource:0}: Error finding container 207f3492f26c495cef6bbffa4030a15f9f871fd2a98f20dbcf685d7a52eceb56: Status 404 returned error can't find the container with id 207f3492f26c495cef6bbffa4030a15f9f871fd2a98f20dbcf685d7a52eceb56 Nov 23 20:23:36 crc kubenswrapper[4726]: E1123 20:23:36.493005 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Nov 23 20:23:36 crc kubenswrapper[4726]: E1123 20:23:36.493047 4726 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Nov 23 20:23:36 crc kubenswrapper[4726]: E1123 20:23:36.493188 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fbb2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(91c39827-8843-4434-8e7f-1ff5be6d9c24): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 23 20:23:36 crc kubenswrapper[4726]: E1123 20:23:36.495025 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="91c39827-8843-4434-8e7f-1ff5be6d9c24" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.795668 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-jnv7b"] Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.798031 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.804208 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.818422 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jnv7b"] Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.883687 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.884500 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbe231ca-6343-4afb-b11b-a74c26a01a2f-config\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.884562 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqswq\" (UniqueName: \"kubernetes.io/projected/fbe231ca-6343-4afb-b11b-a74c26a01a2f-kube-api-access-lqswq\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.884589 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fbe231ca-6343-4afb-b11b-a74c26a01a2f-ovs-rundir\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.884618 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbe231ca-6343-4afb-b11b-a74c26a01a2f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.884647 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe231ca-6343-4afb-b11b-a74c26a01a2f-combined-ca-bundle\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.884671 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fbe231ca-6343-4afb-b11b-a74c26a01a2f-ovn-rundir\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.902099 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.984975 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-config\") pod \"12c8453d-da78-4c44-a770-f1546072da33\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.985729 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n64h\" (UniqueName: \"kubernetes.io/projected/12c8453d-da78-4c44-a770-f1546072da33-kube-api-access-9n64h\") pod \"12c8453d-da78-4c44-a770-f1546072da33\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.987027 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-dns-svc\") pod \"12c8453d-da78-4c44-a770-f1546072da33\" (UID: \"12c8453d-da78-4c44-a770-f1546072da33\") " Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.987118 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac6696-5fc8-4278-8a76-33acaf607ab4-config\") pod \"42ac6696-5fc8-4278-8a76-33acaf607ab4\" (UID: \"42ac6696-5fc8-4278-8a76-33acaf607ab4\") " Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.987247 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28pmq\" (UniqueName: \"kubernetes.io/projected/42ac6696-5fc8-4278-8a76-33acaf607ab4-kube-api-access-28pmq\") pod \"42ac6696-5fc8-4278-8a76-33acaf607ab4\" (UID: \"42ac6696-5fc8-4278-8a76-33acaf607ab4\") " Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.987512 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbe231ca-6343-4afb-b11b-a74c26a01a2f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.987620 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe231ca-6343-4afb-b11b-a74c26a01a2f-combined-ca-bundle\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.988719 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fbe231ca-6343-4afb-b11b-a74c26a01a2f-ovn-rundir\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.991145 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbe231ca-6343-4afb-b11b-a74c26a01a2f-config\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.991503 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqswq\" (UniqueName: \"kubernetes.io/projected/fbe231ca-6343-4afb-b11b-a74c26a01a2f-kube-api-access-lqswq\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.991735 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fbe231ca-6343-4afb-b11b-a74c26a01a2f-ovs-rundir\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.992152 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fbe231ca-6343-4afb-b11b-a74c26a01a2f-ovs-rundir\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.985669 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-config" (OuterVolumeSpecName: "config") pod "12c8453d-da78-4c44-a770-f1546072da33" (UID: "12c8453d-da78-4c44-a770-f1546072da33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.988812 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42ac6696-5fc8-4278-8a76-33acaf607ab4-config" (OuterVolumeSpecName: "config") pod "42ac6696-5fc8-4278-8a76-33acaf607ab4" (UID: "42ac6696-5fc8-4278-8a76-33acaf607ab4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.989250 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "12c8453d-da78-4c44-a770-f1546072da33" (UID: "12c8453d-da78-4c44-a770-f1546072da33"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.994979 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c8453d-da78-4c44-a770-f1546072da33-kube-api-access-9n64h" (OuterVolumeSpecName: "kube-api-access-9n64h") pod "12c8453d-da78-4c44-a770-f1546072da33" (UID: "12c8453d-da78-4c44-a770-f1546072da33"). InnerVolumeSpecName "kube-api-access-9n64h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.995705 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fbe231ca-6343-4afb-b11b-a74c26a01a2f-ovn-rundir\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:36 crc kubenswrapper[4726]: I1123 20:23:36.997554 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbe231ca-6343-4afb-b11b-a74c26a01a2f-config\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.090392 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42ac6696-5fc8-4278-8a76-33acaf607ab4-kube-api-access-28pmq" (OuterVolumeSpecName: "kube-api-access-28pmq") pod "42ac6696-5fc8-4278-8a76-33acaf607ab4" (UID: "42ac6696-5fc8-4278-8a76-33acaf607ab4"). InnerVolumeSpecName "kube-api-access-28pmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.091637 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe231ca-6343-4afb-b11b-a74c26a01a2f-combined-ca-bundle\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.092081 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqswq\" (UniqueName: \"kubernetes.io/projected/fbe231ca-6343-4afb-b11b-a74c26a01a2f-kube-api-access-lqswq\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.093143 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28pmq\" (UniqueName: \"kubernetes.io/projected/42ac6696-5fc8-4278-8a76-33acaf607ab4-kube-api-access-28pmq\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.093162 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.093172 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n64h\" (UniqueName: \"kubernetes.io/projected/12c8453d-da78-4c44-a770-f1546072da33-kube-api-access-9n64h\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.093198 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12c8453d-da78-4c44-a770-f1546072da33-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.093210 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac6696-5fc8-4278-8a76-33acaf607ab4-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.093449 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbe231ca-6343-4afb-b11b-a74c26a01a2f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jnv7b\" (UID: \"fbe231ca-6343-4afb-b11b-a74c26a01a2f\") " pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.222919 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jnv7b" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.229841 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-khj6k"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.264251 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-8wwt7"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.266043 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.272774 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-8wwt7"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.275409 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.298280 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-config\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.298351 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.298423 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.298447 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k6ng\" (UniqueName: \"kubernetes.io/projected/6c41174e-8a3d-4b8d-94c7-83e2447464ae-kube-api-access-2k6ng\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.383256 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1cb13d69-d5a0-4a7b-929b-ed077760725a","Type":"ContainerStarted","Data":"207f3492f26c495cef6bbffa4030a15f9f871fd2a98f20dbcf685d7a52eceb56"} Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.384903 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3647d686-8ad6-4bad-9ca6-ffbaba9bba97","Type":"ContainerStarted","Data":"3d7ba9c12c7bec910ee540aea829fadab59ba2632bcca643602ba966db006022"} Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.393121 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-ftqwl" event={"ID":"12c8453d-da78-4c44-a770-f1546072da33","Type":"ContainerDied","Data":"f079fb412da40dd2ed1e3ba38132983fb19d685407337d3a49be1ab046e1b796"} Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.393191 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-ftqwl" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.399909 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.400088 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k6ng\" (UniqueName: \"kubernetes.io/projected/6c41174e-8a3d-4b8d-94c7-83e2447464ae-kube-api-access-2k6ng\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.400139 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-config\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.400186 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.401503 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.402058 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.402762 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-config\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.403144 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e88e0fb1-897d-454c-9a23-8187592296e5","Type":"ContainerStarted","Data":"979eff5dd3662d35293e2c285637d17b1467f09fadb05045d322b40751fadf79"} Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.420550 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" event={"ID":"42ac6696-5fc8-4278-8a76-33acaf607ab4","Type":"ContainerDied","Data":"bb640a1fb90b73161195d8f6143f4ab096b5c254dcc108bae74fa70cc85c6e73"} Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.420617 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-v9hsl" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.434646 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k6ng\" (UniqueName: \"kubernetes.io/projected/6c41174e-8a3d-4b8d-94c7-83e2447464ae-kube-api-access-2k6ng\") pod \"dnsmasq-dns-65c9b8d4f7-8wwt7\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: E1123 20:23:37.451943 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="91c39827-8843-4434-8e7f-1ff5be6d9c24" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.565654 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-v9hsl"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.575210 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-rq5b9"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.587380 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-v9hsl"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.611390 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.620573 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-ftqwl"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.646939 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-ftqwl"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.682830 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-ggvz7"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.704808 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.712641 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.770658 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-ggvz7"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.845242 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-config\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.845322 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v86fd\" (UniqueName: \"kubernetes.io/projected/b34343ff-43aa-4518-a7d5-96c3b3678e18-kube-api-access-v86fd\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.845383 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.845419 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.845457 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.925817 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jnv7b"] Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.949447 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.949494 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.949546 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.949573 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-config\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.949641 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v86fd\" (UniqueName: \"kubernetes.io/projected/b34343ff-43aa-4518-a7d5-96c3b3678e18-kube-api-access-v86fd\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.950635 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.950673 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.950972 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.951329 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-config\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:37 crc kubenswrapper[4726]: W1123 20:23:37.956267 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbe231ca_6343_4afb_b11b_a74c26a01a2f.slice/crio-ad00b55a4a42d95eb94f6d73cefe93c56d3af3e2d4f7a067b8485b101f019ef3 WatchSource:0}: Error finding container ad00b55a4a42d95eb94f6d73cefe93c56d3af3e2d4f7a067b8485b101f019ef3: Status 404 returned error can't find the container with id ad00b55a4a42d95eb94f6d73cefe93c56d3af3e2d4f7a067b8485b101f019ef3 Nov 23 20:23:37 crc kubenswrapper[4726]: I1123 20:23:37.973749 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v86fd\" (UniqueName: \"kubernetes.io/projected/b34343ff-43aa-4518-a7d5-96c3b3678e18-kube-api-access-v86fd\") pod \"dnsmasq-dns-5c476d78c5-ggvz7\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.096842 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.230656 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.252032 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.359717 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-config\") pod \"31640635-e6ea-4445-bf46-007d16c78378\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.359804 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwt6z\" (UniqueName: \"kubernetes.io/projected/d2275f49-16fd-463d-9d90-314581001af4-kube-api-access-wwt6z\") pod \"d2275f49-16fd-463d-9d90-314581001af4\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.359830 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr996\" (UniqueName: \"kubernetes.io/projected/31640635-e6ea-4445-bf46-007d16c78378-kube-api-access-jr996\") pod \"31640635-e6ea-4445-bf46-007d16c78378\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.359897 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-dns-svc\") pod \"d2275f49-16fd-463d-9d90-314581001af4\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.359959 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-config\") pod \"d2275f49-16fd-463d-9d90-314581001af4\" (UID: \"d2275f49-16fd-463d-9d90-314581001af4\") " Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.360008 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-dns-svc\") pod \"31640635-e6ea-4445-bf46-007d16c78378\" (UID: \"31640635-e6ea-4445-bf46-007d16c78378\") " Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.360933 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-config" (OuterVolumeSpecName: "config") pod "31640635-e6ea-4445-bf46-007d16c78378" (UID: "31640635-e6ea-4445-bf46-007d16c78378"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.361001 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d2275f49-16fd-463d-9d90-314581001af4" (UID: "d2275f49-16fd-463d-9d90-314581001af4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.361081 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-config" (OuterVolumeSpecName: "config") pod "d2275f49-16fd-463d-9d90-314581001af4" (UID: "d2275f49-16fd-463d-9d90-314581001af4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.361159 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "31640635-e6ea-4445-bf46-007d16c78378" (UID: "31640635-e6ea-4445-bf46-007d16c78378"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.371000 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31640635-e6ea-4445-bf46-007d16c78378-kube-api-access-jr996" (OuterVolumeSpecName: "kube-api-access-jr996") pod "31640635-e6ea-4445-bf46-007d16c78378" (UID: "31640635-e6ea-4445-bf46-007d16c78378"). InnerVolumeSpecName "kube-api-access-jr996". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.372058 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2275f49-16fd-463d-9d90-314581001af4-kube-api-access-wwt6z" (OuterVolumeSpecName: "kube-api-access-wwt6z") pod "d2275f49-16fd-463d-9d90-314581001af4" (UID: "d2275f49-16fd-463d-9d90-314581001af4"). InnerVolumeSpecName "kube-api-access-wwt6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.439743 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-khj6k" event={"ID":"d2275f49-16fd-463d-9d90-314581001af4","Type":"ContainerDied","Data":"375157895b3950b8b64a35fbd4aa3b296f69b8babfcec8c8c00bb9c5b86226f9"} Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.439819 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-khj6k" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.447679 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6","Type":"ContainerStarted","Data":"23e09746982c7ccb78dd8daed1438ea3180103f7bbf3840ead65e6402c29f107"} Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.448263 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-8wwt7"] Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.465000 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.465030 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwt6z\" (UniqueName: \"kubernetes.io/projected/d2275f49-16fd-463d-9d90-314581001af4-kube-api-access-wwt6z\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.465040 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr996\" (UniqueName: \"kubernetes.io/projected/31640635-e6ea-4445-bf46-007d16c78378-kube-api-access-jr996\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.465049 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.465058 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2275f49-16fd-463d-9d90-314581001af4-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.465065 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31640635-e6ea-4445-bf46-007d16c78378-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.482314 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1879312-c440-415f-a376-4ab58b99e21d","Type":"ContainerStarted","Data":"2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f"} Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.495973 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" event={"ID":"31640635-e6ea-4445-bf46-007d16c78378","Type":"ContainerDied","Data":"0b26a12bfb1b211628f6d29b8d1f66e70e21f2ff436f7e45e2cd0ab1c075b997"} Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.496056 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-rq5b9" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.499917 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jnv7b" event={"ID":"fbe231ca-6343-4afb-b11b-a74c26a01a2f","Type":"ContainerStarted","Data":"ad00b55a4a42d95eb94f6d73cefe93c56d3af3e2d4f7a067b8485b101f019ef3"} Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.650715 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c8453d-da78-4c44-a770-f1546072da33" path="/var/lib/kubelet/pods/12c8453d-da78-4c44-a770-f1546072da33/volumes" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.654429 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42ac6696-5fc8-4278-8a76-33acaf607ab4" path="/var/lib/kubelet/pods/42ac6696-5fc8-4278-8a76-33acaf607ab4/volumes" Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.654803 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-khj6k"] Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.654824 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-khj6k"] Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.668430 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-ggvz7"] Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.732831 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-rq5b9"] Nov 23 20:23:38 crc kubenswrapper[4726]: I1123 20:23:38.748153 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-rq5b9"] Nov 23 20:23:39 crc kubenswrapper[4726]: I1123 20:23:39.042796 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:23:39 crc kubenswrapper[4726]: I1123 20:23:39.042841 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:23:39 crc kubenswrapper[4726]: I1123 20:23:39.042892 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:23:39 crc kubenswrapper[4726]: I1123 20:23:39.043573 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e0e5405a0e8a49ddd9351d6065e06af6eb07dcd086934b09caf4e017ae67cf3f"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:23:39 crc kubenswrapper[4726]: I1123 20:23:39.043756 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://e0e5405a0e8a49ddd9351d6065e06af6eb07dcd086934b09caf4e017ae67cf3f" gracePeriod=600 Nov 23 20:23:39 crc kubenswrapper[4726]: I1123 20:23:39.514662 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="e0e5405a0e8a49ddd9351d6065e06af6eb07dcd086934b09caf4e017ae67cf3f" exitCode=0 Nov 23 20:23:39 crc kubenswrapper[4726]: I1123 20:23:39.514748 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"e0e5405a0e8a49ddd9351d6065e06af6eb07dcd086934b09caf4e017ae67cf3f"} Nov 23 20:23:39 crc kubenswrapper[4726]: I1123 20:23:39.514802 4726 scope.go:117] "RemoveContainer" containerID="8ef166c526ceeb4bedeeddb339c120f70a8a458484a2d19fd740d49251ee0d41" Nov 23 20:23:40 crc kubenswrapper[4726]: I1123 20:23:40.613668 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31640635-e6ea-4445-bf46-007d16c78378" path="/var/lib/kubelet/pods/31640635-e6ea-4445-bf46-007d16c78378/volumes" Nov 23 20:23:40 crc kubenswrapper[4726]: I1123 20:23:40.615536 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2275f49-16fd-463d-9d90-314581001af4" path="/var/lib/kubelet/pods/d2275f49-16fd-463d-9d90-314581001af4/volumes" Nov 23 20:23:40 crc kubenswrapper[4726]: W1123 20:23:40.627649 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb34343ff_43aa_4518_a7d5_96c3b3678e18.slice/crio-417f9d4ebc33f7b0a7392dd7c3e45f26a9a184437e2be6500d2017fd44303225 WatchSource:0}: Error finding container 417f9d4ebc33f7b0a7392dd7c3e45f26a9a184437e2be6500d2017fd44303225: Status 404 returned error can't find the container with id 417f9d4ebc33f7b0a7392dd7c3e45f26a9a184437e2be6500d2017fd44303225 Nov 23 20:23:40 crc kubenswrapper[4726]: W1123 20:23:40.845448 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c41174e_8a3d_4b8d_94c7_83e2447464ae.slice/crio-c4c6119c5f0fcac876c302564c38bf2f8c1d21f8017f24d5a71e76ec4bc410e7 WatchSource:0}: Error finding container c4c6119c5f0fcac876c302564c38bf2f8c1d21f8017f24d5a71e76ec4bc410e7: Status 404 returned error can't find the container with id c4c6119c5f0fcac876c302564c38bf2f8c1d21f8017f24d5a71e76ec4bc410e7 Nov 23 20:23:41 crc kubenswrapper[4726]: I1123 20:23:41.566930 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1cb13d69-d5a0-4a7b-929b-ed077760725a","Type":"ContainerStarted","Data":"69c356333dab96bcff4426028fe450c51336a7d544fbf8c96588e6a6f026eff5"} Nov 23 20:23:41 crc kubenswrapper[4726]: I1123 20:23:41.569650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" event={"ID":"6c41174e-8a3d-4b8d-94c7-83e2447464ae","Type":"ContainerStarted","Data":"c4c6119c5f0fcac876c302564c38bf2f8c1d21f8017f24d5a71e76ec4bc410e7"} Nov 23 20:23:41 crc kubenswrapper[4726]: I1123 20:23:41.570705 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" event={"ID":"b34343ff-43aa-4518-a7d5-96c3b3678e18","Type":"ContainerStarted","Data":"417f9d4ebc33f7b0a7392dd7c3e45f26a9a184437e2be6500d2017fd44303225"} Nov 23 20:23:41 crc kubenswrapper[4726]: I1123 20:23:41.574013 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"93d2133e107e131190f68c969bc423a889dae2dbf2757948997b4402e77bc885"} Nov 23 20:23:41 crc kubenswrapper[4726]: I1123 20:23:41.578635 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"319fe004-3d9c-4bb9-bb8e-114c8949bcc7","Type":"ContainerStarted","Data":"4d1a590faf2c6e5d6d5e83790d66e63b76ee7b8cef0c1d848259d4a886fab98f"} Nov 23 20:23:41 crc kubenswrapper[4726]: I1123 20:23:41.579155 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 23 20:23:41 crc kubenswrapper[4726]: I1123 20:23:41.580843 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7","Type":"ContainerStarted","Data":"5b2443126f83a13738fd2cdb6081d4753b2dd4566da12fb0752724819dba7da6"} Nov 23 20:23:41 crc kubenswrapper[4726]: I1123 20:23:41.615280 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.954925388 podStartE2EDuration="34.615261173s" podCreationTimestamp="2025-11-23 20:23:07 +0000 UTC" firstStartedPulling="2025-11-23 20:23:08.55667958 +0000 UTC m=+896.705720536" lastFinishedPulling="2025-11-23 20:23:41.217015365 +0000 UTC m=+929.366056321" observedRunningTime="2025-11-23 20:23:41.609275477 +0000 UTC m=+929.758316433" watchObservedRunningTime="2025-11-23 20:23:41.615261173 +0000 UTC m=+929.764302129" Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.598453 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-657sc" event={"ID":"4cb17e51-7e2f-4012-96fd-5f6bd424429c","Type":"ContainerStarted","Data":"15842b3fb3fef1204d89faea09436f81a4b8a17ff9d411b0dca7b0d4e6c5ac01"} Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.600591 4726 generic.go:334] "Generic (PLEG): container finished" podID="e88e0fb1-897d-454c-9a23-8187592296e5" containerID="979eff5dd3662d35293e2c285637d17b1467f09fadb05045d322b40751fadf79" exitCode=0 Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.600719 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e88e0fb1-897d-454c-9a23-8187592296e5","Type":"ContainerDied","Data":"979eff5dd3662d35293e2c285637d17b1467f09fadb05045d322b40751fadf79"} Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.603796 4726 generic.go:334] "Generic (PLEG): container finished" podID="ac53da75-5263-4d97-aa49-b4e8b345c854" containerID="00ea93b732320a353b52ab2fea39ce0672aa8381c0f2ca40325167474c435120" exitCode=0 Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.603878 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sq7pg" event={"ID":"ac53da75-5263-4d97-aa49-b4e8b345c854","Type":"ContainerDied","Data":"00ea93b732320a353b52ab2fea39ce0672aa8381c0f2ca40325167474c435120"} Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.606664 4726 generic.go:334] "Generic (PLEG): container finished" podID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" containerID="0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155" exitCode=0 Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.606773 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" event={"ID":"6c41174e-8a3d-4b8d-94c7-83e2447464ae","Type":"ContainerDied","Data":"0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155"} Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.609386 4726 generic.go:334] "Generic (PLEG): container finished" podID="3647d686-8ad6-4bad-9ca6-ffbaba9bba97" containerID="3d7ba9c12c7bec910ee540aea829fadab59ba2632bcca643602ba966db006022" exitCode=0 Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.609434 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3647d686-8ad6-4bad-9ca6-ffbaba9bba97","Type":"ContainerDied","Data":"3d7ba9c12c7bec910ee540aea829fadab59ba2632bcca643602ba966db006022"} Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.613594 4726 generic.go:334] "Generic (PLEG): container finished" podID="b34343ff-43aa-4518-a7d5-96c3b3678e18" containerID="41f66e41cadc1443f27db47b7b26814ec13f82f321a96c7e2d7939b05e435e3f" exitCode=0 Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.614040 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" event={"ID":"b34343ff-43aa-4518-a7d5-96c3b3678e18","Type":"ContainerDied","Data":"41f66e41cadc1443f27db47b7b26814ec13f82f321a96c7e2d7939b05e435e3f"} Nov 23 20:23:42 crc kubenswrapper[4726]: I1123 20:23:42.801108 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-657sc" podStartSLOduration=19.433716412 podStartE2EDuration="29.80108648s" podCreationTimestamp="2025-11-23 20:23:13 +0000 UTC" firstStartedPulling="2025-11-23 20:23:30.833721846 +0000 UTC m=+918.982762822" lastFinishedPulling="2025-11-23 20:23:41.201091934 +0000 UTC m=+929.350132890" observedRunningTime="2025-11-23 20:23:42.793385936 +0000 UTC m=+930.942426902" watchObservedRunningTime="2025-11-23 20:23:42.80108648 +0000 UTC m=+930.950127436" Nov 23 20:23:43 crc kubenswrapper[4726]: I1123 20:23:43.621427 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-657sc" Nov 23 20:23:47 crc kubenswrapper[4726]: I1123 20:23:47.908084 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.688617 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e88e0fb1-897d-454c-9a23-8187592296e5","Type":"ContainerStarted","Data":"bd8e3b17e491fddf389590c0b98c10911db8399b9b3c3743af42728751c28226"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.692828 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ff67ad55-4c63-45f3-b6f5-12f941bdd2a7","Type":"ContainerStarted","Data":"bdb1321bf648190b924339e2b244d2284430cee6052f6beb4e5cb56d699cdaf0"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.696087 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sq7pg" event={"ID":"ac53da75-5263-4d97-aa49-b4e8b345c854","Type":"ContainerStarted","Data":"67ed03a0aa3250524f971ad77fe86e143a72537941081e79a63964e8d46247b0"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.696124 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sq7pg" event={"ID":"ac53da75-5263-4d97-aa49-b4e8b345c854","Type":"ContainerStarted","Data":"7561045e01f8e5514b64262508d4b0eda48fe1f47cb016eacdbf8af60e64817d"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.696556 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.697752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1cb13d69-d5a0-4a7b-929b-ed077760725a","Type":"ContainerStarted","Data":"17da7aa435db260e59b07c5c5220c1f9e23c5c32e053341239a6150ec4a5f1ab"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.700510 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" event={"ID":"6c41174e-8a3d-4b8d-94c7-83e2447464ae","Type":"ContainerStarted","Data":"faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.700634 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.702844 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3647d686-8ad6-4bad-9ca6-ffbaba9bba97","Type":"ContainerStarted","Data":"07447e154628789b510ad2c091685fbb29cc052a9d7e724fd252a5cffaf5adc1"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.705215 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" event={"ID":"b34343ff-43aa-4518-a7d5-96c3b3678e18","Type":"ContainerStarted","Data":"c66c311edcc2cbb5a34ff94d0ba78eb88b732a23d5e098970d620185822c64f0"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.706733 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jnv7b" event={"ID":"fbe231ca-6343-4afb-b11b-a74c26a01a2f","Type":"ContainerStarted","Data":"217cb9db6710b7c38a91a2538632f341ae6f74abd3e2d1aba72b8e98f500cb86"} Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.726980 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=15.506615893 podStartE2EDuration="42.726957307s" podCreationTimestamp="2025-11-23 20:23:06 +0000 UTC" firstStartedPulling="2025-11-23 20:23:08.682552953 +0000 UTC m=+896.831593909" lastFinishedPulling="2025-11-23 20:23:35.902894367 +0000 UTC m=+924.051935323" observedRunningTime="2025-11-23 20:23:48.720026265 +0000 UTC m=+936.869067221" watchObservedRunningTime="2025-11-23 20:23:48.726957307 +0000 UTC m=+936.875998263" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.752379 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-jnv7b" podStartSLOduration=2.956241231 podStartE2EDuration="12.75235957s" podCreationTimestamp="2025-11-23 20:23:36 +0000 UTC" firstStartedPulling="2025-11-23 20:23:37.959705005 +0000 UTC m=+926.108745961" lastFinishedPulling="2025-11-23 20:23:47.755823344 +0000 UTC m=+935.904864300" observedRunningTime="2025-11-23 20:23:48.745447589 +0000 UTC m=+936.894488565" watchObservedRunningTime="2025-11-23 20:23:48.75235957 +0000 UTC m=+936.901400536" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.787597 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-sq7pg" podStartSLOduration=30.192587339 podStartE2EDuration="35.787579855s" podCreationTimestamp="2025-11-23 20:23:13 +0000 UTC" firstStartedPulling="2025-11-23 20:23:35.606992693 +0000 UTC m=+923.756033649" lastFinishedPulling="2025-11-23 20:23:41.201985199 +0000 UTC m=+929.351026165" observedRunningTime="2025-11-23 20:23:48.785481687 +0000 UTC m=+936.934522663" watchObservedRunningTime="2025-11-23 20:23:48.787579855 +0000 UTC m=+936.936620821" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.821377 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" podStartSLOduration=11.415292575 podStartE2EDuration="11.82136062s" podCreationTimestamp="2025-11-23 20:23:37 +0000 UTC" firstStartedPulling="2025-11-23 20:23:40.88532954 +0000 UTC m=+929.034370516" lastFinishedPulling="2025-11-23 20:23:41.291397605 +0000 UTC m=+929.440438561" observedRunningTime="2025-11-23 20:23:48.81484362 +0000 UTC m=+936.963884596" watchObservedRunningTime="2025-11-23 20:23:48.82136062 +0000 UTC m=+936.970401566" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.839160 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=21.655558181 podStartE2EDuration="32.839136882s" podCreationTimestamp="2025-11-23 20:23:16 +0000 UTC" firstStartedPulling="2025-11-23 20:23:36.49916738 +0000 UTC m=+924.648208336" lastFinishedPulling="2025-11-23 20:23:47.682746091 +0000 UTC m=+935.831787037" observedRunningTime="2025-11-23 20:23:48.836335134 +0000 UTC m=+936.985376100" watchObservedRunningTime="2025-11-23 20:23:48.839136882 +0000 UTC m=+936.988177838" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.860931 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" podStartSLOduration=11.276290026 podStartE2EDuration="11.860909735s" podCreationTimestamp="2025-11-23 20:23:37 +0000 UTC" firstStartedPulling="2025-11-23 20:23:40.6354288 +0000 UTC m=+928.784469756" lastFinishedPulling="2025-11-23 20:23:41.220048509 +0000 UTC m=+929.369089465" observedRunningTime="2025-11-23 20:23:48.860434102 +0000 UTC m=+937.009475058" watchObservedRunningTime="2025-11-23 20:23:48.860909735 +0000 UTC m=+937.009950711" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.879776 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.495726126 podStartE2EDuration="44.879752537s" podCreationTimestamp="2025-11-23 20:23:04 +0000 UTC" firstStartedPulling="2025-11-23 20:23:08.084061018 +0000 UTC m=+896.233101974" lastFinishedPulling="2025-11-23 20:23:36.468087429 +0000 UTC m=+924.617128385" observedRunningTime="2025-11-23 20:23:48.87697372 +0000 UTC m=+937.026014686" watchObservedRunningTime="2025-11-23 20:23:48.879752537 +0000 UTC m=+937.028793493" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.894963 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.946365385 podStartE2EDuration="35.894945077s" podCreationTimestamp="2025-11-23 20:23:13 +0000 UTC" firstStartedPulling="2025-11-23 20:23:30.851384755 +0000 UTC m=+919.000425751" lastFinishedPulling="2025-11-23 20:23:47.799964487 +0000 UTC m=+935.949005443" observedRunningTime="2025-11-23 20:23:48.893354253 +0000 UTC m=+937.042395229" watchObservedRunningTime="2025-11-23 20:23:48.894945077 +0000 UTC m=+937.043986033" Nov 23 20:23:48 crc kubenswrapper[4726]: I1123 20:23:48.946653 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:23:49 crc kubenswrapper[4726]: I1123 20:23:49.613714 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:49 crc kubenswrapper[4726]: I1123 20:23:49.714395 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:50 crc kubenswrapper[4726]: I1123 20:23:50.451183 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:50 crc kubenswrapper[4726]: I1123 20:23:50.614313 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:50 crc kubenswrapper[4726]: I1123 20:23:50.630487 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:50 crc kubenswrapper[4726]: I1123 20:23:50.678995 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:50 crc kubenswrapper[4726]: I1123 20:23:50.728830 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:50 crc kubenswrapper[4726]: I1123 20:23:50.792712 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 23 20:23:50 crc kubenswrapper[4726]: I1123 20:23:50.798882 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.101529 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.103243 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.105634 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.105753 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.105761 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-whbxj" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.106099 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.142605 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.235853 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.235944 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhpgx\" (UniqueName: \"kubernetes.io/projected/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-kube-api-access-fhpgx\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.236014 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.236041 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-scripts\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.236110 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.236148 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.236199 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-config\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.337350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.337404 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-scripts\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.337472 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.337512 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.337545 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-config\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.337578 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.337611 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhpgx\" (UniqueName: \"kubernetes.io/projected/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-kube-api-access-fhpgx\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.338526 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-config\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.338542 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-scripts\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.338904 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.343831 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.344272 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.345811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.360637 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhpgx\" (UniqueName: \"kubernetes.io/projected/3d5bc7b6-0493-40ff-808b-6d13bb31c8c0-kube-api-access-fhpgx\") pod \"ovn-northd-0\" (UID: \"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0\") " pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.421268 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.733249 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"91c39827-8843-4434-8e7f-1ff5be6d9c24","Type":"ContainerStarted","Data":"afda28590e38150743d2f7734882f8429d718e8fbd2782bf68a9341cdabe347a"} Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.734364 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.892195 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.265376521 podStartE2EDuration="42.892168205s" podCreationTimestamp="2025-11-23 20:23:09 +0000 UTC" firstStartedPulling="2025-11-23 20:23:11.100510374 +0000 UTC m=+899.249551330" lastFinishedPulling="2025-11-23 20:23:50.727302058 +0000 UTC m=+938.876343014" observedRunningTime="2025-11-23 20:23:51.753596648 +0000 UTC m=+939.902637604" watchObservedRunningTime="2025-11-23 20:23:51.892168205 +0000 UTC m=+940.041209161" Nov 23 20:23:51 crc kubenswrapper[4726]: I1123 20:23:51.895680 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 23 20:23:51 crc kubenswrapper[4726]: W1123 20:23:51.902007 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d5bc7b6_0493_40ff_808b_6d13bb31c8c0.slice/crio-ca16dccf8b729ce88a260fe5a30db677e2cb76670419f316631c4c6e55117611 WatchSource:0}: Error finding container ca16dccf8b729ce88a260fe5a30db677e2cb76670419f316631c4c6e55117611: Status 404 returned error can't find the container with id ca16dccf8b729ce88a260fe5a30db677e2cb76670419f316631c4c6e55117611 Nov 23 20:23:52 crc kubenswrapper[4726]: I1123 20:23:52.617593 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:52 crc kubenswrapper[4726]: I1123 20:23:52.747488 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0","Type":"ContainerStarted","Data":"ca16dccf8b729ce88a260fe5a30db677e2cb76670419f316631c4c6e55117611"} Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.100107 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.162272 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-8wwt7"] Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.162499 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" podUID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" containerName="dnsmasq-dns" containerID="cri-o://faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952" gracePeriod=10 Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.556846 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.687245 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-dns-svc\") pod \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.687357 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-ovsdbserver-sb\") pod \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.687393 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-config\") pod \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.687540 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k6ng\" (UniqueName: \"kubernetes.io/projected/6c41174e-8a3d-4b8d-94c7-83e2447464ae-kube-api-access-2k6ng\") pod \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\" (UID: \"6c41174e-8a3d-4b8d-94c7-83e2447464ae\") " Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.698398 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c41174e-8a3d-4b8d-94c7-83e2447464ae-kube-api-access-2k6ng" (OuterVolumeSpecName: "kube-api-access-2k6ng") pod "6c41174e-8a3d-4b8d-94c7-83e2447464ae" (UID: "6c41174e-8a3d-4b8d-94c7-83e2447464ae"). InnerVolumeSpecName "kube-api-access-2k6ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.723896 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-config" (OuterVolumeSpecName: "config") pod "6c41174e-8a3d-4b8d-94c7-83e2447464ae" (UID: "6c41174e-8a3d-4b8d-94c7-83e2447464ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.724413 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6c41174e-8a3d-4b8d-94c7-83e2447464ae" (UID: "6c41174e-8a3d-4b8d-94c7-83e2447464ae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.727468 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c41174e-8a3d-4b8d-94c7-83e2447464ae" (UID: "6c41174e-8a3d-4b8d-94c7-83e2447464ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.756600 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0","Type":"ContainerStarted","Data":"42ea241ba575f024b5ddf345954284f93c492b7199b05ba34b54c9bad3ee19ff"} Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.756654 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d5bc7b6-0493-40ff-808b-6d13bb31c8c0","Type":"ContainerStarted","Data":"207b4b51e1caf6c2a94b05d6449fdcfcd36f8230d6913da6e67fe207f077a883"} Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.758516 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.758968 4726 generic.go:334] "Generic (PLEG): container finished" podID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" containerID="faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952" exitCode=0 Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.759003 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" event={"ID":"6c41174e-8a3d-4b8d-94c7-83e2447464ae","Type":"ContainerDied","Data":"faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952"} Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.759024 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" event={"ID":"6c41174e-8a3d-4b8d-94c7-83e2447464ae","Type":"ContainerDied","Data":"c4c6119c5f0fcac876c302564c38bf2f8c1d21f8017f24d5a71e76ec4bc410e7"} Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.759044 4726 scope.go:117] "RemoveContainer" containerID="faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.759969 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-8wwt7" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.776594 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.674411426 podStartE2EDuration="2.776570556s" podCreationTimestamp="2025-11-23 20:23:51 +0000 UTC" firstStartedPulling="2025-11-23 20:23:51.907699165 +0000 UTC m=+940.056740121" lastFinishedPulling="2025-11-23 20:23:53.009858295 +0000 UTC m=+941.158899251" observedRunningTime="2025-11-23 20:23:53.775784755 +0000 UTC m=+941.924825721" watchObservedRunningTime="2025-11-23 20:23:53.776570556 +0000 UTC m=+941.925611532" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.789369 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.789403 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.789415 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c41174e-8a3d-4b8d-94c7-83e2447464ae-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.789424 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k6ng\" (UniqueName: \"kubernetes.io/projected/6c41174e-8a3d-4b8d-94c7-83e2447464ae-kube-api-access-2k6ng\") on node \"crc\" DevicePath \"\"" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.790285 4726 scope.go:117] "RemoveContainer" containerID="0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.814226 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-8wwt7"] Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.814974 4726 scope.go:117] "RemoveContainer" containerID="faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952" Nov 23 20:23:53 crc kubenswrapper[4726]: E1123 20:23:53.816760 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952\": container with ID starting with faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952 not found: ID does not exist" containerID="faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.816974 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952"} err="failed to get container status \"faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952\": rpc error: code = NotFound desc = could not find container \"faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952\": container with ID starting with faa916859954e79848eb13267d900209f11a58d881b48a7d38d6ad456503d952 not found: ID does not exist" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.817131 4726 scope.go:117] "RemoveContainer" containerID="0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155" Nov 23 20:23:53 crc kubenswrapper[4726]: E1123 20:23:53.817608 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155\": container with ID starting with 0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155 not found: ID does not exist" containerID="0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.817666 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155"} err="failed to get container status \"0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155\": rpc error: code = NotFound desc = could not find container \"0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155\": container with ID starting with 0d993bb776b90df9f726759d6c6b27e7f0cdeb0ae04d36ac57b0198625ca8155 not found: ID does not exist" Nov 23 20:23:53 crc kubenswrapper[4726]: I1123 20:23:53.823247 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-8wwt7"] Nov 23 20:23:54 crc kubenswrapper[4726]: E1123 20:23:54.154975 4726 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.110:52108->38.102.83.110:43095: write tcp 38.102.83.110:52108->38.102.83.110:43095: write: broken pipe Nov 23 20:23:54 crc kubenswrapper[4726]: I1123 20:23:54.603816 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" path="/var/lib/kubelet/pods/6c41174e-8a3d-4b8d-94c7-83e2447464ae/volumes" Nov 23 20:23:56 crc kubenswrapper[4726]: I1123 20:23:56.193149 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 23 20:23:56 crc kubenswrapper[4726]: I1123 20:23:56.193505 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 23 20:23:56 crc kubenswrapper[4726]: I1123 20:23:56.297837 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 23 20:23:56 crc kubenswrapper[4726]: I1123 20:23:56.863130 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.263037 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4d43-account-create-jvb7s"] Nov 23 20:23:57 crc kubenswrapper[4726]: E1123 20:23:57.263375 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" containerName="dnsmasq-dns" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.263387 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" containerName="dnsmasq-dns" Nov 23 20:23:57 crc kubenswrapper[4726]: E1123 20:23:57.263413 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" containerName="init" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.263419 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" containerName="init" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.263607 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c41174e-8a3d-4b8d-94c7-83e2447464ae" containerName="dnsmasq-dns" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.264111 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.266785 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.306401 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-4k65v"] Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.308357 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4k65v" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.321441 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4k65v"] Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.326025 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4d43-account-create-jvb7s"] Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.344993 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-operator-scripts\") pod \"keystone-db-create-4k65v\" (UID: \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\") " pod="openstack/keystone-db-create-4k65v" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.345053 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7t2w\" (UniqueName: \"kubernetes.io/projected/3fd3ee85-5046-4edf-9d13-413525f0752e-kube-api-access-z7t2w\") pod \"keystone-4d43-account-create-jvb7s\" (UID: \"3fd3ee85-5046-4edf-9d13-413525f0752e\") " pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.345112 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz2cw\" (UniqueName: \"kubernetes.io/projected/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-kube-api-access-vz2cw\") pod \"keystone-db-create-4k65v\" (UID: \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\") " pod="openstack/keystone-db-create-4k65v" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.345141 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fd3ee85-5046-4edf-9d13-413525f0752e-operator-scripts\") pod \"keystone-4d43-account-create-jvb7s\" (UID: \"3fd3ee85-5046-4edf-9d13-413525f0752e\") " pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.447157 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-operator-scripts\") pod \"keystone-db-create-4k65v\" (UID: \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\") " pod="openstack/keystone-db-create-4k65v" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.447223 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7t2w\" (UniqueName: \"kubernetes.io/projected/3fd3ee85-5046-4edf-9d13-413525f0752e-kube-api-access-z7t2w\") pod \"keystone-4d43-account-create-jvb7s\" (UID: \"3fd3ee85-5046-4edf-9d13-413525f0752e\") " pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.447293 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz2cw\" (UniqueName: \"kubernetes.io/projected/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-kube-api-access-vz2cw\") pod \"keystone-db-create-4k65v\" (UID: \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\") " pod="openstack/keystone-db-create-4k65v" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.447319 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fd3ee85-5046-4edf-9d13-413525f0752e-operator-scripts\") pod \"keystone-4d43-account-create-jvb7s\" (UID: \"3fd3ee85-5046-4edf-9d13-413525f0752e\") " pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.447979 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-operator-scripts\") pod \"keystone-db-create-4k65v\" (UID: \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\") " pod="openstack/keystone-db-create-4k65v" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.448050 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fd3ee85-5046-4edf-9d13-413525f0752e-operator-scripts\") pod \"keystone-4d43-account-create-jvb7s\" (UID: \"3fd3ee85-5046-4edf-9d13-413525f0752e\") " pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.465539 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7t2w\" (UniqueName: \"kubernetes.io/projected/3fd3ee85-5046-4edf-9d13-413525f0752e-kube-api-access-z7t2w\") pod \"keystone-4d43-account-create-jvb7s\" (UID: \"3fd3ee85-5046-4edf-9d13-413525f0752e\") " pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.465982 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz2cw\" (UniqueName: \"kubernetes.io/projected/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-kube-api-access-vz2cw\") pod \"keystone-db-create-4k65v\" (UID: \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\") " pod="openstack/keystone-db-create-4k65v" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.531509 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-x7cvm"] Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.532542 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x7cvm" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.541515 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-x7cvm"] Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.549102 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eee8014b-ecc7-4756-967e-3da08ee9192e-operator-scripts\") pod \"placement-db-create-x7cvm\" (UID: \"eee8014b-ecc7-4756-967e-3da08ee9192e\") " pod="openstack/placement-db-create-x7cvm" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.549176 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cwvl\" (UniqueName: \"kubernetes.io/projected/eee8014b-ecc7-4756-967e-3da08ee9192e-kube-api-access-6cwvl\") pod \"placement-db-create-x7cvm\" (UID: \"eee8014b-ecc7-4756-967e-3da08ee9192e\") " pod="openstack/placement-db-create-x7cvm" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.586665 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.626788 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4k65v" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.631044 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-118c-account-create-jmbld"] Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.632136 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.636037 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.650889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eee8014b-ecc7-4756-967e-3da08ee9192e-operator-scripts\") pod \"placement-db-create-x7cvm\" (UID: \"eee8014b-ecc7-4756-967e-3da08ee9192e\") " pod="openstack/placement-db-create-x7cvm" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.650946 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfqdp\" (UniqueName: \"kubernetes.io/projected/6643a95f-f556-4e3e-9a45-faab4f4772b6-kube-api-access-tfqdp\") pod \"placement-118c-account-create-jmbld\" (UID: \"6643a95f-f556-4e3e-9a45-faab4f4772b6\") " pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.650988 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cwvl\" (UniqueName: \"kubernetes.io/projected/eee8014b-ecc7-4756-967e-3da08ee9192e-kube-api-access-6cwvl\") pod \"placement-db-create-x7cvm\" (UID: \"eee8014b-ecc7-4756-967e-3da08ee9192e\") " pod="openstack/placement-db-create-x7cvm" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.651033 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6643a95f-f556-4e3e-9a45-faab4f4772b6-operator-scripts\") pod \"placement-118c-account-create-jmbld\" (UID: \"6643a95f-f556-4e3e-9a45-faab4f4772b6\") " pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.651629 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eee8014b-ecc7-4756-967e-3da08ee9192e-operator-scripts\") pod \"placement-db-create-x7cvm\" (UID: \"eee8014b-ecc7-4756-967e-3da08ee9192e\") " pod="openstack/placement-db-create-x7cvm" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.652982 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-118c-account-create-jmbld"] Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.665955 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cwvl\" (UniqueName: \"kubernetes.io/projected/eee8014b-ecc7-4756-967e-3da08ee9192e-kube-api-access-6cwvl\") pod \"placement-db-create-x7cvm\" (UID: \"eee8014b-ecc7-4756-967e-3da08ee9192e\") " pod="openstack/placement-db-create-x7cvm" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.752691 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfqdp\" (UniqueName: \"kubernetes.io/projected/6643a95f-f556-4e3e-9a45-faab4f4772b6-kube-api-access-tfqdp\") pod \"placement-118c-account-create-jmbld\" (UID: \"6643a95f-f556-4e3e-9a45-faab4f4772b6\") " pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.753051 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6643a95f-f556-4e3e-9a45-faab4f4772b6-operator-scripts\") pod \"placement-118c-account-create-jmbld\" (UID: \"6643a95f-f556-4e3e-9a45-faab4f4772b6\") " pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.754000 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6643a95f-f556-4e3e-9a45-faab4f4772b6-operator-scripts\") pod \"placement-118c-account-create-jmbld\" (UID: \"6643a95f-f556-4e3e-9a45-faab4f4772b6\") " pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.770088 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfqdp\" (UniqueName: \"kubernetes.io/projected/6643a95f-f556-4e3e-9a45-faab4f4772b6-kube-api-access-tfqdp\") pod \"placement-118c-account-create-jmbld\" (UID: \"6643a95f-f556-4e3e-9a45-faab4f4772b6\") " pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.865489 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x7cvm" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.956169 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:57 crc kubenswrapper[4726]: I1123 20:23:57.956218 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.019796 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.051049 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4k65v"] Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.057242 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4d43-account-create-jvb7s"] Nov 23 20:23:58 crc kubenswrapper[4726]: W1123 20:23:58.065244 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeba1f02c_d02a_4c20_a09f_b3292cd23a0d.slice/crio-3fb90ce627afa50eedd6c39f6a4b813e56ddf4a269912cc3f8d393148501d66d WatchSource:0}: Error finding container 3fb90ce627afa50eedd6c39f6a4b813e56ddf4a269912cc3f8d393148501d66d: Status 404 returned error can't find the container with id 3fb90ce627afa50eedd6c39f6a4b813e56ddf4a269912cc3f8d393148501d66d Nov 23 20:23:58 crc kubenswrapper[4726]: W1123 20:23:58.067766 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fd3ee85_5046_4edf_9d13_413525f0752e.slice/crio-30da328d492f4fb4172ea99e737abb45c4facd46db782259534037c4b1d9267f WatchSource:0}: Error finding container 30da328d492f4fb4172ea99e737abb45c4facd46db782259534037c4b1d9267f: Status 404 returned error can't find the container with id 30da328d492f4fb4172ea99e737abb45c4facd46db782259534037c4b1d9267f Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.107584 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.175656 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-eacb-account-create-wn7kh"] Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.177069 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.179480 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.189058 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-d6nsj"] Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.190202 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d6nsj" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.217973 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-eacb-account-create-wn7kh"] Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.234670 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-d6nsj"] Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.356273 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-x7cvm"] Nov 23 20:23:58 crc kubenswrapper[4726]: W1123 20:23:58.360426 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeee8014b_ecc7_4756_967e_3da08ee9192e.slice/crio-c0bfee4f9106e1c4e0b2674a37620d919039d457e57b40bdba8ad4dfe69ac784 WatchSource:0}: Error finding container c0bfee4f9106e1c4e0b2674a37620d919039d457e57b40bdba8ad4dfe69ac784: Status 404 returned error can't find the container with id c0bfee4f9106e1c4e0b2674a37620d919039d457e57b40bdba8ad4dfe69ac784 Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.366248 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0dacd12-8219-4c35-8d8b-3903cfb98c63-operator-scripts\") pod \"glance-db-create-d6nsj\" (UID: \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\") " pod="openstack/glance-db-create-d6nsj" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.366319 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7sgx\" (UniqueName: \"kubernetes.io/projected/7c50815e-46f6-4f70-be21-4c2c9f98bf35-kube-api-access-c7sgx\") pod \"glance-eacb-account-create-wn7kh\" (UID: \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\") " pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.366342 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c50815e-46f6-4f70-be21-4c2c9f98bf35-operator-scripts\") pod \"glance-eacb-account-create-wn7kh\" (UID: \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\") " pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.366397 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj4hk\" (UniqueName: \"kubernetes.io/projected/e0dacd12-8219-4c35-8d8b-3903cfb98c63-kube-api-access-gj4hk\") pod \"glance-db-create-d6nsj\" (UID: \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\") " pod="openstack/glance-db-create-d6nsj" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.467437 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7sgx\" (UniqueName: \"kubernetes.io/projected/7c50815e-46f6-4f70-be21-4c2c9f98bf35-kube-api-access-c7sgx\") pod \"glance-eacb-account-create-wn7kh\" (UID: \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\") " pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.467481 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c50815e-46f6-4f70-be21-4c2c9f98bf35-operator-scripts\") pod \"glance-eacb-account-create-wn7kh\" (UID: \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\") " pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.467545 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj4hk\" (UniqueName: \"kubernetes.io/projected/e0dacd12-8219-4c35-8d8b-3903cfb98c63-kube-api-access-gj4hk\") pod \"glance-db-create-d6nsj\" (UID: \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\") " pod="openstack/glance-db-create-d6nsj" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.467585 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0dacd12-8219-4c35-8d8b-3903cfb98c63-operator-scripts\") pod \"glance-db-create-d6nsj\" (UID: \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\") " pod="openstack/glance-db-create-d6nsj" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.468292 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0dacd12-8219-4c35-8d8b-3903cfb98c63-operator-scripts\") pod \"glance-db-create-d6nsj\" (UID: \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\") " pod="openstack/glance-db-create-d6nsj" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.469147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c50815e-46f6-4f70-be21-4c2c9f98bf35-operator-scripts\") pod \"glance-eacb-account-create-wn7kh\" (UID: \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\") " pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.486290 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj4hk\" (UniqueName: \"kubernetes.io/projected/e0dacd12-8219-4c35-8d8b-3903cfb98c63-kube-api-access-gj4hk\") pod \"glance-db-create-d6nsj\" (UID: \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\") " pod="openstack/glance-db-create-d6nsj" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.486475 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7sgx\" (UniqueName: \"kubernetes.io/projected/7c50815e-46f6-4f70-be21-4c2c9f98bf35-kube-api-access-c7sgx\") pod \"glance-eacb-account-create-wn7kh\" (UID: \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\") " pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.513701 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.567549 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-118c-account-create-jmbld"] Nov 23 20:23:58 crc kubenswrapper[4726]: W1123 20:23:58.577834 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6643a95f_f556_4e3e_9a45_faab4f4772b6.slice/crio-f59bf3a088fde5187d570a49c55192333e6d6e880bf0c8a967ec6a626d6e4360 WatchSource:0}: Error finding container f59bf3a088fde5187d570a49c55192333e6d6e880bf0c8a967ec6a626d6e4360: Status 404 returned error can't find the container with id f59bf3a088fde5187d570a49c55192333e6d6e880bf0c8a967ec6a626d6e4360 Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.599834 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d6nsj" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.801473 4726 generic.go:334] "Generic (PLEG): container finished" podID="3fd3ee85-5046-4edf-9d13-413525f0752e" containerID="c6853c4b4d49170bfa1661ced9b933946d1faccf6560802174cb7ed1ebf3e480" exitCode=0 Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.801610 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4d43-account-create-jvb7s" event={"ID":"3fd3ee85-5046-4edf-9d13-413525f0752e","Type":"ContainerDied","Data":"c6853c4b4d49170bfa1661ced9b933946d1faccf6560802174cb7ed1ebf3e480"} Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.801657 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4d43-account-create-jvb7s" event={"ID":"3fd3ee85-5046-4edf-9d13-413525f0752e","Type":"ContainerStarted","Data":"30da328d492f4fb4172ea99e737abb45c4facd46db782259534037c4b1d9267f"} Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.804900 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-118c-account-create-jmbld" event={"ID":"6643a95f-f556-4e3e-9a45-faab4f4772b6","Type":"ContainerStarted","Data":"f59bf3a088fde5187d570a49c55192333e6d6e880bf0c8a967ec6a626d6e4360"} Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.806374 4726 generic.go:334] "Generic (PLEG): container finished" podID="eba1f02c-d02a-4c20-a09f-b3292cd23a0d" containerID="861df19309cc381e13b50ead4bc23c6863d43f94befe38d31db2119f945c3c27" exitCode=0 Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.806437 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4k65v" event={"ID":"eba1f02c-d02a-4c20-a09f-b3292cd23a0d","Type":"ContainerDied","Data":"861df19309cc381e13b50ead4bc23c6863d43f94befe38d31db2119f945c3c27"} Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.806459 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4k65v" event={"ID":"eba1f02c-d02a-4c20-a09f-b3292cd23a0d","Type":"ContainerStarted","Data":"3fb90ce627afa50eedd6c39f6a4b813e56ddf4a269912cc3f8d393148501d66d"} Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.808098 4726 generic.go:334] "Generic (PLEG): container finished" podID="eee8014b-ecc7-4756-967e-3da08ee9192e" containerID="5b12cec731b054532248b400473ac69038e166fd7a10d4df07e78851dc36ecda" exitCode=0 Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.808314 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x7cvm" event={"ID":"eee8014b-ecc7-4756-967e-3da08ee9192e","Type":"ContainerDied","Data":"5b12cec731b054532248b400473ac69038e166fd7a10d4df07e78851dc36ecda"} Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.808340 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x7cvm" event={"ID":"eee8014b-ecc7-4756-967e-3da08ee9192e","Type":"ContainerStarted","Data":"c0bfee4f9106e1c4e0b2674a37620d919039d457e57b40bdba8ad4dfe69ac784"} Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.907086 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 23 20:23:58 crc kubenswrapper[4726]: I1123 20:23:58.978141 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-eacb-account-create-wn7kh"] Nov 23 20:23:58 crc kubenswrapper[4726]: W1123 20:23:58.981482 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c50815e_46f6_4f70_be21_4c2c9f98bf35.slice/crio-c60fa83c7ca3e5e42ad5059859b23a9a5375e28a85659f5a41fc743c0829b03e WatchSource:0}: Error finding container c60fa83c7ca3e5e42ad5059859b23a9a5375e28a85659f5a41fc743c0829b03e: Status 404 returned error can't find the container with id c60fa83c7ca3e5e42ad5059859b23a9a5375e28a85659f5a41fc743c0829b03e Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.133347 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-d6nsj"] Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.821489 4726 generic.go:334] "Generic (PLEG): container finished" podID="6643a95f-f556-4e3e-9a45-faab4f4772b6" containerID="b88d48fea82f8ade138ebc58abe6c0feac1624b7f966409096d9e0d3985823a1" exitCode=0 Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.821613 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-118c-account-create-jmbld" event={"ID":"6643a95f-f556-4e3e-9a45-faab4f4772b6","Type":"ContainerDied","Data":"b88d48fea82f8ade138ebc58abe6c0feac1624b7f966409096d9e0d3985823a1"} Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.823984 4726 generic.go:334] "Generic (PLEG): container finished" podID="7c50815e-46f6-4f70-be21-4c2c9f98bf35" containerID="c6514513a7733b2cf84485c1c6ceea1579f973052e8a50517556f92a36fcffda" exitCode=0 Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.824092 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eacb-account-create-wn7kh" event={"ID":"7c50815e-46f6-4f70-be21-4c2c9f98bf35","Type":"ContainerDied","Data":"c6514513a7733b2cf84485c1c6ceea1579f973052e8a50517556f92a36fcffda"} Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.824130 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eacb-account-create-wn7kh" event={"ID":"7c50815e-46f6-4f70-be21-4c2c9f98bf35","Type":"ContainerStarted","Data":"c60fa83c7ca3e5e42ad5059859b23a9a5375e28a85659f5a41fc743c0829b03e"} Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.827188 4726 generic.go:334] "Generic (PLEG): container finished" podID="e0dacd12-8219-4c35-8d8b-3903cfb98c63" containerID="4cd02756ae106783e889092006a66e574deeb783f7245b06539f4e0e2909f8f0" exitCode=0 Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.827587 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d6nsj" event={"ID":"e0dacd12-8219-4c35-8d8b-3903cfb98c63","Type":"ContainerDied","Data":"4cd02756ae106783e889092006a66e574deeb783f7245b06539f4e0e2909f8f0"} Nov 23 20:23:59 crc kubenswrapper[4726]: I1123 20:23:59.827634 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d6nsj" event={"ID":"e0dacd12-8219-4c35-8d8b-3903cfb98c63","Type":"ContainerStarted","Data":"f81090e3d5f7d6834f6d4a02dc956617a40e44736d5fd4728d1e1521f53e8ecb"} Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.092312 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.258727 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x7cvm" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.340861 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4k65v" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.354584 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.401362 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eee8014b-ecc7-4756-967e-3da08ee9192e-operator-scripts\") pod \"eee8014b-ecc7-4756-967e-3da08ee9192e\" (UID: \"eee8014b-ecc7-4756-967e-3da08ee9192e\") " Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.401575 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cwvl\" (UniqueName: \"kubernetes.io/projected/eee8014b-ecc7-4756-967e-3da08ee9192e-kube-api-access-6cwvl\") pod \"eee8014b-ecc7-4756-967e-3da08ee9192e\" (UID: \"eee8014b-ecc7-4756-967e-3da08ee9192e\") " Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.401902 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eee8014b-ecc7-4756-967e-3da08ee9192e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eee8014b-ecc7-4756-967e-3da08ee9192e" (UID: "eee8014b-ecc7-4756-967e-3da08ee9192e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.407898 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eee8014b-ecc7-4756-967e-3da08ee9192e-kube-api-access-6cwvl" (OuterVolumeSpecName: "kube-api-access-6cwvl") pod "eee8014b-ecc7-4756-967e-3da08ee9192e" (UID: "eee8014b-ecc7-4756-967e-3da08ee9192e"). InnerVolumeSpecName "kube-api-access-6cwvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.502774 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fd3ee85-5046-4edf-9d13-413525f0752e-operator-scripts\") pod \"3fd3ee85-5046-4edf-9d13-413525f0752e\" (UID: \"3fd3ee85-5046-4edf-9d13-413525f0752e\") " Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.502835 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz2cw\" (UniqueName: \"kubernetes.io/projected/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-kube-api-access-vz2cw\") pod \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\" (UID: \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\") " Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.502961 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7t2w\" (UniqueName: \"kubernetes.io/projected/3fd3ee85-5046-4edf-9d13-413525f0752e-kube-api-access-z7t2w\") pod \"3fd3ee85-5046-4edf-9d13-413525f0752e\" (UID: \"3fd3ee85-5046-4edf-9d13-413525f0752e\") " Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.503072 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-operator-scripts\") pod \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\" (UID: \"eba1f02c-d02a-4c20-a09f-b3292cd23a0d\") " Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.503182 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fd3ee85-5046-4edf-9d13-413525f0752e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3fd3ee85-5046-4edf-9d13-413525f0752e" (UID: "3fd3ee85-5046-4edf-9d13-413525f0752e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.503370 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cwvl\" (UniqueName: \"kubernetes.io/projected/eee8014b-ecc7-4756-967e-3da08ee9192e-kube-api-access-6cwvl\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.503390 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fd3ee85-5046-4edf-9d13-413525f0752e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.503401 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eee8014b-ecc7-4756-967e-3da08ee9192e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.503627 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eba1f02c-d02a-4c20-a09f-b3292cd23a0d" (UID: "eba1f02c-d02a-4c20-a09f-b3292cd23a0d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.506046 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-kube-api-access-vz2cw" (OuterVolumeSpecName: "kube-api-access-vz2cw") pod "eba1f02c-d02a-4c20-a09f-b3292cd23a0d" (UID: "eba1f02c-d02a-4c20-a09f-b3292cd23a0d"). InnerVolumeSpecName "kube-api-access-vz2cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.506225 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fd3ee85-5046-4edf-9d13-413525f0752e-kube-api-access-z7t2w" (OuterVolumeSpecName: "kube-api-access-z7t2w") pod "3fd3ee85-5046-4edf-9d13-413525f0752e" (UID: "3fd3ee85-5046-4edf-9d13-413525f0752e"). InnerVolumeSpecName "kube-api-access-z7t2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.604666 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz2cw\" (UniqueName: \"kubernetes.io/projected/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-kube-api-access-vz2cw\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.604693 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7t2w\" (UniqueName: \"kubernetes.io/projected/3fd3ee85-5046-4edf-9d13-413525f0752e-kube-api-access-z7t2w\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.604704 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eba1f02c-d02a-4c20-a09f-b3292cd23a0d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.837613 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4k65v" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.837625 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4k65v" event={"ID":"eba1f02c-d02a-4c20-a09f-b3292cd23a0d","Type":"ContainerDied","Data":"3fb90ce627afa50eedd6c39f6a4b813e56ddf4a269912cc3f8d393148501d66d"} Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.837668 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fb90ce627afa50eedd6c39f6a4b813e56ddf4a269912cc3f8d393148501d66d" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.840477 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x7cvm" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.840553 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x7cvm" event={"ID":"eee8014b-ecc7-4756-967e-3da08ee9192e","Type":"ContainerDied","Data":"c0bfee4f9106e1c4e0b2674a37620d919039d457e57b40bdba8ad4dfe69ac784"} Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.840600 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0bfee4f9106e1c4e0b2674a37620d919039d457e57b40bdba8ad4dfe69ac784" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.844951 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4d43-account-create-jvb7s" event={"ID":"3fd3ee85-5046-4edf-9d13-413525f0752e","Type":"ContainerDied","Data":"30da328d492f4fb4172ea99e737abb45c4facd46db782259534037c4b1d9267f"} Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.844993 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30da328d492f4fb4172ea99e737abb45c4facd46db782259534037c4b1d9267f" Nov 23 20:24:00 crc kubenswrapper[4726]: I1123 20:24:00.845089 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4d43-account-create-jvb7s" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.143688 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d6nsj" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.284971 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.313449 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.318040 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0dacd12-8219-4c35-8d8b-3903cfb98c63-operator-scripts\") pod \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\" (UID: \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\") " Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.318109 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj4hk\" (UniqueName: \"kubernetes.io/projected/e0dacd12-8219-4c35-8d8b-3903cfb98c63-kube-api-access-gj4hk\") pod \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\" (UID: \"e0dacd12-8219-4c35-8d8b-3903cfb98c63\") " Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.319014 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0dacd12-8219-4c35-8d8b-3903cfb98c63-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e0dacd12-8219-4c35-8d8b-3903cfb98c63" (UID: "e0dacd12-8219-4c35-8d8b-3903cfb98c63"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.323694 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0dacd12-8219-4c35-8d8b-3903cfb98c63-kube-api-access-gj4hk" (OuterVolumeSpecName: "kube-api-access-gj4hk") pod "e0dacd12-8219-4c35-8d8b-3903cfb98c63" (UID: "e0dacd12-8219-4c35-8d8b-3903cfb98c63"). InnerVolumeSpecName "kube-api-access-gj4hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.419279 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c50815e-46f6-4f70-be21-4c2c9f98bf35-operator-scripts\") pod \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\" (UID: \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\") " Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.419430 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfqdp\" (UniqueName: \"kubernetes.io/projected/6643a95f-f556-4e3e-9a45-faab4f4772b6-kube-api-access-tfqdp\") pod \"6643a95f-f556-4e3e-9a45-faab4f4772b6\" (UID: \"6643a95f-f556-4e3e-9a45-faab4f4772b6\") " Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.419493 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7sgx\" (UniqueName: \"kubernetes.io/projected/7c50815e-46f6-4f70-be21-4c2c9f98bf35-kube-api-access-c7sgx\") pod \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\" (UID: \"7c50815e-46f6-4f70-be21-4c2c9f98bf35\") " Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.419574 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6643a95f-f556-4e3e-9a45-faab4f4772b6-operator-scripts\") pod \"6643a95f-f556-4e3e-9a45-faab4f4772b6\" (UID: \"6643a95f-f556-4e3e-9a45-faab4f4772b6\") " Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.419732 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c50815e-46f6-4f70-be21-4c2c9f98bf35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c50815e-46f6-4f70-be21-4c2c9f98bf35" (UID: "7c50815e-46f6-4f70-be21-4c2c9f98bf35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.420089 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6643a95f-f556-4e3e-9a45-faab4f4772b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6643a95f-f556-4e3e-9a45-faab4f4772b6" (UID: "6643a95f-f556-4e3e-9a45-faab4f4772b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.420152 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj4hk\" (UniqueName: \"kubernetes.io/projected/e0dacd12-8219-4c35-8d8b-3903cfb98c63-kube-api-access-gj4hk\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.420172 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c50815e-46f6-4f70-be21-4c2c9f98bf35-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.420185 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0dacd12-8219-4c35-8d8b-3903cfb98c63-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.422898 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6643a95f-f556-4e3e-9a45-faab4f4772b6-kube-api-access-tfqdp" (OuterVolumeSpecName: "kube-api-access-tfqdp") pod "6643a95f-f556-4e3e-9a45-faab4f4772b6" (UID: "6643a95f-f556-4e3e-9a45-faab4f4772b6"). InnerVolumeSpecName "kube-api-access-tfqdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.423164 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c50815e-46f6-4f70-be21-4c2c9f98bf35-kube-api-access-c7sgx" (OuterVolumeSpecName: "kube-api-access-c7sgx") pod "7c50815e-46f6-4f70-be21-4c2c9f98bf35" (UID: "7c50815e-46f6-4f70-be21-4c2c9f98bf35"). InnerVolumeSpecName "kube-api-access-c7sgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.522049 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfqdp\" (UniqueName: \"kubernetes.io/projected/6643a95f-f556-4e3e-9a45-faab4f4772b6-kube-api-access-tfqdp\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.522087 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7sgx\" (UniqueName: \"kubernetes.io/projected/7c50815e-46f6-4f70-be21-4c2c9f98bf35-kube-api-access-c7sgx\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.522099 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6643a95f-f556-4e3e-9a45-faab4f4772b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.856918 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-118c-account-create-jmbld" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.856929 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-118c-account-create-jmbld" event={"ID":"6643a95f-f556-4e3e-9a45-faab4f4772b6","Type":"ContainerDied","Data":"f59bf3a088fde5187d570a49c55192333e6d6e880bf0c8a967ec6a626d6e4360"} Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.857384 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f59bf3a088fde5187d570a49c55192333e6d6e880bf0c8a967ec6a626d6e4360" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.858948 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eacb-account-create-wn7kh" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.858983 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eacb-account-create-wn7kh" event={"ID":"7c50815e-46f6-4f70-be21-4c2c9f98bf35","Type":"ContainerDied","Data":"c60fa83c7ca3e5e42ad5059859b23a9a5375e28a85659f5a41fc743c0829b03e"} Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.859042 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c60fa83c7ca3e5e42ad5059859b23a9a5375e28a85659f5a41fc743c0829b03e" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.861130 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d6nsj" event={"ID":"e0dacd12-8219-4c35-8d8b-3903cfb98c63","Type":"ContainerDied","Data":"f81090e3d5f7d6834f6d4a02dc956617a40e44736d5fd4728d1e1521f53e8ecb"} Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.861193 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f81090e3d5f7d6834f6d4a02dc956617a40e44736d5fd4728d1e1521f53e8ecb" Nov 23 20:24:01 crc kubenswrapper[4726]: I1123 20:24:01.861441 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d6nsj" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369053 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-fffk2"] Nov 23 20:24:03 crc kubenswrapper[4726]: E1123 20:24:03.369371 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee8014b-ecc7-4756-967e-3da08ee9192e" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369383 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee8014b-ecc7-4756-967e-3da08ee9192e" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: E1123 20:24:03.369401 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0dacd12-8219-4c35-8d8b-3903cfb98c63" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369407 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0dacd12-8219-4c35-8d8b-3903cfb98c63" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: E1123 20:24:03.369416 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c50815e-46f6-4f70-be21-4c2c9f98bf35" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369424 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c50815e-46f6-4f70-be21-4c2c9f98bf35" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: E1123 20:24:03.369432 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba1f02c-d02a-4c20-a09f-b3292cd23a0d" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369438 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba1f02c-d02a-4c20-a09f-b3292cd23a0d" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: E1123 20:24:03.369453 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd3ee85-5046-4edf-9d13-413525f0752e" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369458 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd3ee85-5046-4edf-9d13-413525f0752e" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: E1123 20:24:03.369474 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6643a95f-f556-4e3e-9a45-faab4f4772b6" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369480 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6643a95f-f556-4e3e-9a45-faab4f4772b6" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369619 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="eba1f02c-d02a-4c20-a09f-b3292cd23a0d" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369632 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee8014b-ecc7-4756-967e-3da08ee9192e" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369639 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0dacd12-8219-4c35-8d8b-3903cfb98c63" containerName="mariadb-database-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369645 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c50815e-46f6-4f70-be21-4c2c9f98bf35" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369652 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fd3ee85-5046-4edf-9d13-413525f0752e" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.369664 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6643a95f-f556-4e3e-9a45-faab4f4772b6" containerName="mariadb-account-create" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.370177 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.372346 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kppfj" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.377061 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.380191 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6lt2\" (UniqueName: \"kubernetes.io/projected/6877e1a3-06c3-4619-9ec1-61cecf78c66f-kube-api-access-m6lt2\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.380261 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-config-data\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.380310 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-combined-ca-bundle\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.380352 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-db-sync-config-data\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.394077 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fffk2"] Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.481491 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-db-sync-config-data\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.481568 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6lt2\" (UniqueName: \"kubernetes.io/projected/6877e1a3-06c3-4619-9ec1-61cecf78c66f-kube-api-access-m6lt2\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.481625 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-config-data\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.481644 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-combined-ca-bundle\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.486248 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-db-sync-config-data\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.486472 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-combined-ca-bundle\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.498942 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6lt2\" (UniqueName: \"kubernetes.io/projected/6877e1a3-06c3-4619-9ec1-61cecf78c66f-kube-api-access-m6lt2\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.499806 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-config-data\") pod \"glance-db-sync-fffk2\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:03 crc kubenswrapper[4726]: I1123 20:24:03.684307 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:04 crc kubenswrapper[4726]: I1123 20:24:04.221617 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fffk2"] Nov 23 20:24:04 crc kubenswrapper[4726]: I1123 20:24:04.898161 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fffk2" event={"ID":"6877e1a3-06c3-4619-9ec1-61cecf78c66f","Type":"ContainerStarted","Data":"bb6fbbf4cef4e5a1126f1ecbcd0449c38542b1ae1be6a1fd5830de3633d8acd4"} Nov 23 20:24:06 crc kubenswrapper[4726]: I1123 20:24:06.483046 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 23 20:24:09 crc kubenswrapper[4726]: I1123 20:24:09.965900 4726 generic.go:334] "Generic (PLEG): container finished" podID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerID="23e09746982c7ccb78dd8daed1438ea3180103f7bbf3840ead65e6402c29f107" exitCode=0 Nov 23 20:24:09 crc kubenswrapper[4726]: I1123 20:24:09.965928 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6","Type":"ContainerDied","Data":"23e09746982c7ccb78dd8daed1438ea3180103f7bbf3840ead65e6402c29f107"} Nov 23 20:24:09 crc kubenswrapper[4726]: I1123 20:24:09.971936 4726 generic.go:334] "Generic (PLEG): container finished" podID="c1879312-c440-415f-a376-4ab58b99e21d" containerID="2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f" exitCode=0 Nov 23 20:24:09 crc kubenswrapper[4726]: I1123 20:24:09.971988 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1879312-c440-415f-a376-4ab58b99e21d","Type":"ContainerDied","Data":"2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f"} Nov 23 20:24:13 crc kubenswrapper[4726]: I1123 20:24:13.955581 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-657sc" podUID="4cb17e51-7e2f-4012-96fd-5f6bd424429c" containerName="ovn-controller" probeResult="failure" output=< Nov 23 20:24:13 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 23 20:24:13 crc kubenswrapper[4726]: > Nov 23 20:24:16 crc kubenswrapper[4726]: I1123 20:24:16.042824 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1879312-c440-415f-a376-4ab58b99e21d","Type":"ContainerStarted","Data":"1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18"} Nov 23 20:24:16 crc kubenswrapper[4726]: I1123 20:24:16.045641 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 23 20:24:16 crc kubenswrapper[4726]: I1123 20:24:16.046529 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6","Type":"ContainerStarted","Data":"2471650b3dd77658e1c60d02e48d6b6a1f06b40fb3b935a5ed1903e883f8b86c"} Nov 23 20:24:16 crc kubenswrapper[4726]: I1123 20:24:16.047610 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:24:16 crc kubenswrapper[4726]: I1123 20:24:16.084588 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.737356594 podStartE2EDuration="1m13.084572635s" podCreationTimestamp="2025-11-23 20:23:03 +0000 UTC" firstStartedPulling="2025-11-23 20:23:05.552996512 +0000 UTC m=+893.702037468" lastFinishedPulling="2025-11-23 20:23:35.900212563 +0000 UTC m=+924.049253509" observedRunningTime="2025-11-23 20:24:16.079600307 +0000 UTC m=+964.228641253" watchObservedRunningTime="2025-11-23 20:24:16.084572635 +0000 UTC m=+964.233613591" Nov 23 20:24:16 crc kubenswrapper[4726]: I1123 20:24:16.109178 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.698019369 podStartE2EDuration="1m13.109159106s" podCreationTimestamp="2025-11-23 20:23:03 +0000 UTC" firstStartedPulling="2025-11-23 20:23:06.054196786 +0000 UTC m=+894.203237742" lastFinishedPulling="2025-11-23 20:23:36.465336523 +0000 UTC m=+924.614377479" observedRunningTime="2025-11-23 20:24:16.106737599 +0000 UTC m=+964.255778565" watchObservedRunningTime="2025-11-23 20:24:16.109159106 +0000 UTC m=+964.258200062" Nov 23 20:24:17 crc kubenswrapper[4726]: I1123 20:24:17.055136 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fffk2" event={"ID":"6877e1a3-06c3-4619-9ec1-61cecf78c66f","Type":"ContainerStarted","Data":"b4ff4cd21e8d372185cb63ca270970b780134976af3311ba7521402de7aeba88"} Nov 23 20:24:17 crc kubenswrapper[4726]: I1123 20:24:17.075262 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-fffk2" podStartSLOduration=2.536255377 podStartE2EDuration="14.075243808s" podCreationTimestamp="2025-11-23 20:24:03 +0000 UTC" firstStartedPulling="2025-11-23 20:24:04.222313862 +0000 UTC m=+952.371354818" lastFinishedPulling="2025-11-23 20:24:15.761302293 +0000 UTC m=+963.910343249" observedRunningTime="2025-11-23 20:24:17.074260391 +0000 UTC m=+965.223301367" watchObservedRunningTime="2025-11-23 20:24:17.075243808 +0000 UTC m=+965.224284764" Nov 23 20:24:18 crc kubenswrapper[4726]: I1123 20:24:18.958562 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-657sc" podUID="4cb17e51-7e2f-4012-96fd-5f6bd424429c" containerName="ovn-controller" probeResult="failure" output=< Nov 23 20:24:18 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 23 20:24:18 crc kubenswrapper[4726]: > Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.000493 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.001911 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sq7pg" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.242270 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-657sc-config-skzhv"] Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.243331 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.245784 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.307045 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-657sc-config-skzhv"] Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.369738 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.369837 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-additional-scripts\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.369914 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-log-ovn\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.369967 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run-ovn\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.370235 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p66gc\" (UniqueName: \"kubernetes.io/projected/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-kube-api-access-p66gc\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.370290 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-scripts\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.472773 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p66gc\" (UniqueName: \"kubernetes.io/projected/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-kube-api-access-p66gc\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.473644 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-scripts\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.473775 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.474013 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-additional-scripts\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.475434 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-log-ovn\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.475533 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-log-ovn\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.474244 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.475561 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-scripts\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.474666 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-additional-scripts\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.476026 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run-ovn\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.476091 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run-ovn\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.497374 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p66gc\" (UniqueName: \"kubernetes.io/projected/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-kube-api-access-p66gc\") pod \"ovn-controller-657sc-config-skzhv\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:19 crc kubenswrapper[4726]: I1123 20:24:19.565992 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:20 crc kubenswrapper[4726]: W1123 20:24:20.025196 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43029e5c_4a8f_49ce_a56b_c3dcc9155e4d.slice/crio-e749e66059aa160f3d929405ed9cd8be06fb0e48b6c3dc7d5da0ed73db142a33 WatchSource:0}: Error finding container e749e66059aa160f3d929405ed9cd8be06fb0e48b6c3dc7d5da0ed73db142a33: Status 404 returned error can't find the container with id e749e66059aa160f3d929405ed9cd8be06fb0e48b6c3dc7d5da0ed73db142a33 Nov 23 20:24:20 crc kubenswrapper[4726]: I1123 20:24:20.025314 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-657sc-config-skzhv"] Nov 23 20:24:20 crc kubenswrapper[4726]: I1123 20:24:20.082683 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-657sc-config-skzhv" event={"ID":"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d","Type":"ContainerStarted","Data":"e749e66059aa160f3d929405ed9cd8be06fb0e48b6c3dc7d5da0ed73db142a33"} Nov 23 20:24:21 crc kubenswrapper[4726]: I1123 20:24:21.095335 4726 generic.go:334] "Generic (PLEG): container finished" podID="43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" containerID="ea52e905d86a603a499077a3ba92ccddae806f3d27b4e25ca709a5dbda1af821" exitCode=0 Nov 23 20:24:21 crc kubenswrapper[4726]: I1123 20:24:21.095399 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-657sc-config-skzhv" event={"ID":"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d","Type":"ContainerDied","Data":"ea52e905d86a603a499077a3ba92ccddae806f3d27b4e25ca709a5dbda1af821"} Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.456394 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.628176 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run\") pod \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.628344 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-additional-scripts\") pod \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.628338 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run" (OuterVolumeSpecName: "var-run") pod "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" (UID: "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.628420 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-scripts\") pod \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.628463 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p66gc\" (UniqueName: \"kubernetes.io/projected/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-kube-api-access-p66gc\") pod \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.628503 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-log-ovn\") pod \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.628542 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run-ovn\") pod \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\" (UID: \"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d\") " Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.629239 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" (UID: "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.629470 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" (UID: "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.629502 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" (UID: "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.630441 4726 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.630476 4726 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-run\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.630499 4726 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.630511 4726 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.664062 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-scripts" (OuterVolumeSpecName: "scripts") pod "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" (UID: "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.664138 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-kube-api-access-p66gc" (OuterVolumeSpecName: "kube-api-access-p66gc") pod "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" (UID: "43029e5c-4a8f-49ce-a56b-c3dcc9155e4d"). InnerVolumeSpecName "kube-api-access-p66gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.731454 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:22 crc kubenswrapper[4726]: I1123 20:24:22.731491 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p66gc\" (UniqueName: \"kubernetes.io/projected/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d-kube-api-access-p66gc\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.116200 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-657sc-config-skzhv" event={"ID":"43029e5c-4a8f-49ce-a56b-c3dcc9155e4d","Type":"ContainerDied","Data":"e749e66059aa160f3d929405ed9cd8be06fb0e48b6c3dc7d5da0ed73db142a33"} Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.116243 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e749e66059aa160f3d929405ed9cd8be06fb0e48b6c3dc7d5da0ed73db142a33" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.116259 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc-config-skzhv" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.552623 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-657sc-config-skzhv"] Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.562815 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-657sc-config-skzhv"] Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.678748 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-657sc-config-q4cgq"] Nov 23 20:24:23 crc kubenswrapper[4726]: E1123 20:24:23.679086 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" containerName="ovn-config" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.679101 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" containerName="ovn-config" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.679261 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" containerName="ovn-config" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.679921 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.682331 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.737428 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-657sc-config-q4cgq"] Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.856786 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.856883 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run-ovn\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.856933 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk9hb\" (UniqueName: \"kubernetes.io/projected/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-kube-api-access-wk9hb\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.856949 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-scripts\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.856968 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-additional-scripts\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.857065 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-log-ovn\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.952466 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-657sc" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.958769 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.958823 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run-ovn\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.958861 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk9hb\" (UniqueName: \"kubernetes.io/projected/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-kube-api-access-wk9hb\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.958893 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-scripts\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.958908 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-additional-scripts\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.958929 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-log-ovn\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.959217 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-log-ovn\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.959281 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.959359 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run-ovn\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.960066 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-additional-scripts\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:23 crc kubenswrapper[4726]: I1123 20:24:23.961622 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-scripts\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:24 crc kubenswrapper[4726]: I1123 20:24:24.002266 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk9hb\" (UniqueName: \"kubernetes.io/projected/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-kube-api-access-wk9hb\") pod \"ovn-controller-657sc-config-q4cgq\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:24 crc kubenswrapper[4726]: I1123 20:24:24.125907 4726 generic.go:334] "Generic (PLEG): container finished" podID="6877e1a3-06c3-4619-9ec1-61cecf78c66f" containerID="b4ff4cd21e8d372185cb63ca270970b780134976af3311ba7521402de7aeba88" exitCode=0 Nov 23 20:24:24 crc kubenswrapper[4726]: I1123 20:24:24.125958 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fffk2" event={"ID":"6877e1a3-06c3-4619-9ec1-61cecf78c66f","Type":"ContainerDied","Data":"b4ff4cd21e8d372185cb63ca270970b780134976af3311ba7521402de7aeba88"} Nov 23 20:24:24 crc kubenswrapper[4726]: I1123 20:24:24.296052 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:24 crc kubenswrapper[4726]: I1123 20:24:24.598763 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43029e5c-4a8f-49ce-a56b-c3dcc9155e4d" path="/var/lib/kubelet/pods/43029e5c-4a8f-49ce-a56b-c3dcc9155e4d/volumes" Nov 23 20:24:24 crc kubenswrapper[4726]: I1123 20:24:24.819037 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-657sc-config-q4cgq"] Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.134372 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-657sc-config-q4cgq" event={"ID":"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc","Type":"ContainerStarted","Data":"7fab2d3abbe85b46b6e9b4e1c02375e6a1debadb52c63e6cf6e837834c652849"} Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.270383 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.509924 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.687091 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6lt2\" (UniqueName: \"kubernetes.io/projected/6877e1a3-06c3-4619-9ec1-61cecf78c66f-kube-api-access-m6lt2\") pod \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.687145 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-config-data\") pod \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.687191 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-db-sync-config-data\") pod \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.687316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-combined-ca-bundle\") pod \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\" (UID: \"6877e1a3-06c3-4619-9ec1-61cecf78c66f\") " Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.692318 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6877e1a3-06c3-4619-9ec1-61cecf78c66f" (UID: "6877e1a3-06c3-4619-9ec1-61cecf78c66f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.707993 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6877e1a3-06c3-4619-9ec1-61cecf78c66f-kube-api-access-m6lt2" (OuterVolumeSpecName: "kube-api-access-m6lt2") pod "6877e1a3-06c3-4619-9ec1-61cecf78c66f" (UID: "6877e1a3-06c3-4619-9ec1-61cecf78c66f"). InnerVolumeSpecName "kube-api-access-m6lt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.711035 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6877e1a3-06c3-4619-9ec1-61cecf78c66f" (UID: "6877e1a3-06c3-4619-9ec1-61cecf78c66f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.742205 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-config-data" (OuterVolumeSpecName: "config-data") pod "6877e1a3-06c3-4619-9ec1-61cecf78c66f" (UID: "6877e1a3-06c3-4619-9ec1-61cecf78c66f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.789120 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.789339 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.789349 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6lt2\" (UniqueName: \"kubernetes.io/projected/6877e1a3-06c3-4619-9ec1-61cecf78c66f-kube-api-access-m6lt2\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:25 crc kubenswrapper[4726]: I1123 20:24:25.789361 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6877e1a3-06c3-4619-9ec1-61cecf78c66f-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.143223 4726 generic.go:334] "Generic (PLEG): container finished" podID="5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" containerID="aac7e9a0984302acfad53201f5828de638a7de333ad6b6d93667b6b7a76fcad3" exitCode=0 Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.143855 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-657sc-config-q4cgq" event={"ID":"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc","Type":"ContainerDied","Data":"aac7e9a0984302acfad53201f5828de638a7de333ad6b6d93667b6b7a76fcad3"} Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.148606 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fffk2" event={"ID":"6877e1a3-06c3-4619-9ec1-61cecf78c66f","Type":"ContainerDied","Data":"bb6fbbf4cef4e5a1126f1ecbcd0449c38542b1ae1be6a1fd5830de3633d8acd4"} Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.148664 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb6fbbf4cef4e5a1126f1ecbcd0449c38542b1ae1be6a1fd5830de3633d8acd4" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.148748 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fffk2" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.665446 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-569d458467-ffqcl"] Nov 23 20:24:26 crc kubenswrapper[4726]: E1123 20:24:26.665826 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6877e1a3-06c3-4619-9ec1-61cecf78c66f" containerName="glance-db-sync" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.665843 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6877e1a3-06c3-4619-9ec1-61cecf78c66f" containerName="glance-db-sync" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.666017 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6877e1a3-06c3-4619-9ec1-61cecf78c66f" containerName="glance-db-sync" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.666807 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.689067 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-569d458467-ffqcl"] Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.827093 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbj86\" (UniqueName: \"kubernetes.io/projected/862ffc28-0d10-4190-9654-51ae9590043e-kube-api-access-bbj86\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.827143 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-sb\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.827221 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-config\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.827247 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-nb\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.827269 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-dns-svc\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.928863 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbj86\" (UniqueName: \"kubernetes.io/projected/862ffc28-0d10-4190-9654-51ae9590043e-kube-api-access-bbj86\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.929144 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-sb\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.929205 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-config\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.929228 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-nb\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.929249 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-dns-svc\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.930318 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-dns-svc\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.931111 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-sb\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.931677 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-config\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.932187 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-nb\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:26 crc kubenswrapper[4726]: I1123 20:24:26.950122 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbj86\" (UniqueName: \"kubernetes.io/projected/862ffc28-0d10-4190-9654-51ae9590043e-kube-api-access-bbj86\") pod \"dnsmasq-dns-569d458467-ffqcl\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.004374 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:27 crc kubenswrapper[4726]: W1123 20:24:27.446754 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod862ffc28_0d10_4190_9654_51ae9590043e.slice/crio-4b673c4a6931aa78f246965a6250664acfb63a23ceed1f96438f8b9625560eac WatchSource:0}: Error finding container 4b673c4a6931aa78f246965a6250664acfb63a23ceed1f96438f8b9625560eac: Status 404 returned error can't find the container with id 4b673c4a6931aa78f246965a6250664acfb63a23ceed1f96438f8b9625560eac Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.449692 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-569d458467-ffqcl"] Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.473370 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.642758 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-scripts\") pod \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.642824 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-additional-scripts\") pod \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.642946 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run-ovn\") pod \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.642974 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run\") pod \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.643002 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-log-ovn\") pod \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.643122 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk9hb\" (UniqueName: \"kubernetes.io/projected/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-kube-api-access-wk9hb\") pod \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\" (UID: \"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc\") " Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.643334 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run" (OuterVolumeSpecName: "var-run") pod "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" (UID: "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.643664 4726 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.643694 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" (UID: "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.643714 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" (UID: "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.643949 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" (UID: "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.644225 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-scripts" (OuterVolumeSpecName: "scripts") pod "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" (UID: "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.647251 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-kube-api-access-wk9hb" (OuterVolumeSpecName: "kube-api-access-wk9hb") pod "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" (UID: "5cbfd1b5-f9a8-4583-9889-63a13fc6cccc"). InnerVolumeSpecName "kube-api-access-wk9hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.744909 4726 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.745462 4726 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.745479 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk9hb\" (UniqueName: \"kubernetes.io/projected/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-kube-api-access-wk9hb\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.745494 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:27 crc kubenswrapper[4726]: I1123 20:24:27.745507 4726 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.166984 4726 generic.go:334] "Generic (PLEG): container finished" podID="862ffc28-0d10-4190-9654-51ae9590043e" containerID="e728c2dae0ea69563bc7191687530716dfa0097e064bf8a3f753e86f503db4f7" exitCode=0 Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.167283 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-ffqcl" event={"ID":"862ffc28-0d10-4190-9654-51ae9590043e","Type":"ContainerDied","Data":"e728c2dae0ea69563bc7191687530716dfa0097e064bf8a3f753e86f503db4f7"} Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.167349 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-ffqcl" event={"ID":"862ffc28-0d10-4190-9654-51ae9590043e","Type":"ContainerStarted","Data":"4b673c4a6931aa78f246965a6250664acfb63a23ceed1f96438f8b9625560eac"} Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.169817 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-657sc-config-q4cgq" event={"ID":"5cbfd1b5-f9a8-4583-9889-63a13fc6cccc","Type":"ContainerDied","Data":"7fab2d3abbe85b46b6e9b4e1c02375e6a1debadb52c63e6cf6e837834c652849"} Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.169852 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fab2d3abbe85b46b6e9b4e1c02375e6a1debadb52c63e6cf6e837834c652849" Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.169912 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-657sc-config-q4cgq" Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.552100 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-657sc-config-q4cgq"] Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.559915 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-657sc-config-q4cgq"] Nov 23 20:24:28 crc kubenswrapper[4726]: I1123 20:24:28.600544 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" path="/var/lib/kubelet/pods/5cbfd1b5-f9a8-4583-9889-63a13fc6cccc/volumes" Nov 23 20:24:29 crc kubenswrapper[4726]: I1123 20:24:29.199514 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-ffqcl" event={"ID":"862ffc28-0d10-4190-9654-51ae9590043e","Type":"ContainerStarted","Data":"e658d0749861ccc937b6325f5382fbd538c49a561a8aae44ab581690dd14187b"} Nov 23 20:24:29 crc kubenswrapper[4726]: I1123 20:24:29.199697 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:29 crc kubenswrapper[4726]: I1123 20:24:29.225552 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-569d458467-ffqcl" podStartSLOduration=3.225532885 podStartE2EDuration="3.225532885s" podCreationTimestamp="2025-11-23 20:24:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:24:29.219157889 +0000 UTC m=+977.368198865" watchObservedRunningTime="2025-11-23 20:24:29.225532885 +0000 UTC m=+977.374573841" Nov 23 20:24:34 crc kubenswrapper[4726]: I1123 20:24:34.818183 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.157771 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-m4fr4"] Nov 23 20:24:35 crc kubenswrapper[4726]: E1123 20:24:35.158338 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" containerName="ovn-config" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.158360 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" containerName="ovn-config" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.158557 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbfd1b5-f9a8-4583-9889-63a13fc6cccc" containerName="ovn-config" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.159174 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.232792 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-m4fr4"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.270082 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.289921 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-operator-scripts\") pod \"cinder-db-create-m4fr4\" (UID: \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\") " pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.289967 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5sph\" (UniqueName: \"kubernetes.io/projected/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-kube-api-access-n5sph\") pod \"cinder-db-create-m4fr4\" (UID: \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\") " pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.318541 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-x7d9j"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.319646 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.340609 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-x7d9j"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.391806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-operator-scripts\") pod \"cinder-db-create-m4fr4\" (UID: \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\") " pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.391909 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5sph\" (UniqueName: \"kubernetes.io/projected/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-kube-api-access-n5sph\") pod \"cinder-db-create-m4fr4\" (UID: \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\") " pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.392132 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5d0d-account-create-tn2pk"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.392447 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-operator-scripts\") pod \"cinder-db-create-m4fr4\" (UID: \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\") " pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.393470 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.398757 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.431485 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5sph\" (UniqueName: \"kubernetes.io/projected/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-kube-api-access-n5sph\") pod \"cinder-db-create-m4fr4\" (UID: \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\") " pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.449549 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5d0d-account-create-tn2pk"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.475588 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.501774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c2nl\" (UniqueName: \"kubernetes.io/projected/aa225da7-111f-449f-8e6e-d2a73bb39a8f-kube-api-access-7c2nl\") pod \"barbican-db-create-x7d9j\" (UID: \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\") " pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.502122 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbj5s\" (UniqueName: \"kubernetes.io/projected/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-kube-api-access-cbj5s\") pod \"cinder-5d0d-account-create-tn2pk\" (UID: \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\") " pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.502168 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa225da7-111f-449f-8e6e-d2a73bb39a8f-operator-scripts\") pod \"barbican-db-create-x7d9j\" (UID: \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\") " pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.502212 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-operator-scripts\") pod \"cinder-5d0d-account-create-tn2pk\" (UID: \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\") " pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.538532 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b6a0-account-create-2zr8p"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.539512 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.546066 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.567244 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b6a0-account-create-2zr8p"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.607395 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c2nl\" (UniqueName: \"kubernetes.io/projected/aa225da7-111f-449f-8e6e-d2a73bb39a8f-kube-api-access-7c2nl\") pod \"barbican-db-create-x7d9j\" (UID: \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\") " pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.607460 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbj5s\" (UniqueName: \"kubernetes.io/projected/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-kube-api-access-cbj5s\") pod \"cinder-5d0d-account-create-tn2pk\" (UID: \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\") " pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.607505 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa225da7-111f-449f-8e6e-d2a73bb39a8f-operator-scripts\") pod \"barbican-db-create-x7d9j\" (UID: \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\") " pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.607558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-operator-scripts\") pod \"cinder-5d0d-account-create-tn2pk\" (UID: \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\") " pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.608306 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-operator-scripts\") pod \"cinder-5d0d-account-create-tn2pk\" (UID: \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\") " pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.608789 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa225da7-111f-449f-8e6e-d2a73bb39a8f-operator-scripts\") pod \"barbican-db-create-x7d9j\" (UID: \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\") " pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.647640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c2nl\" (UniqueName: \"kubernetes.io/projected/aa225da7-111f-449f-8e6e-d2a73bb39a8f-kube-api-access-7c2nl\") pod \"barbican-db-create-x7d9j\" (UID: \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\") " pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.685495 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbj5s\" (UniqueName: \"kubernetes.io/projected/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-kube-api-access-cbj5s\") pod \"cinder-5d0d-account-create-tn2pk\" (UID: \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\") " pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.706341 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.708594 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c555255-c599-4b91-9a76-c860a2c99228-operator-scripts\") pod \"barbican-b6a0-account-create-2zr8p\" (UID: \"1c555255-c599-4b91-9a76-c860a2c99228\") " pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.708667 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ddv4\" (UniqueName: \"kubernetes.io/projected/1c555255-c599-4b91-9a76-c860a2c99228-kube-api-access-7ddv4\") pod \"barbican-b6a0-account-create-2zr8p\" (UID: \"1c555255-c599-4b91-9a76-c860a2c99228\") " pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.769969 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-psjjc"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.771118 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.784365 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-psjjc"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.815701 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ddv4\" (UniqueName: \"kubernetes.io/projected/1c555255-c599-4b91-9a76-c860a2c99228-kube-api-access-7ddv4\") pod \"barbican-b6a0-account-create-2zr8p\" (UID: \"1c555255-c599-4b91-9a76-c860a2c99228\") " pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.815847 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c555255-c599-4b91-9a76-c860a2c99228-operator-scripts\") pod \"barbican-b6a0-account-create-2zr8p\" (UID: \"1c555255-c599-4b91-9a76-c860a2c99228\") " pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.816544 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c555255-c599-4b91-9a76-c860a2c99228-operator-scripts\") pod \"barbican-b6a0-account-create-2zr8p\" (UID: \"1c555255-c599-4b91-9a76-c860a2c99228\") " pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.863566 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ddv4\" (UniqueName: \"kubernetes.io/projected/1c555255-c599-4b91-9a76-c860a2c99228-kube-api-access-7ddv4\") pod \"barbican-b6a0-account-create-2zr8p\" (UID: \"1c555255-c599-4b91-9a76-c860a2c99228\") " pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.918402 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-769a-account-create-qj28b"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.919663 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57msx\" (UniqueName: \"kubernetes.io/projected/5c555bb2-ca8a-412e-852f-8f928feeab8d-kube-api-access-57msx\") pod \"neutron-db-create-psjjc\" (UID: \"5c555bb2-ca8a-412e-852f-8f928feeab8d\") " pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.919760 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c555bb2-ca8a-412e-852f-8f928feeab8d-operator-scripts\") pod \"neutron-db-create-psjjc\" (UID: \"5c555bb2-ca8a-412e-852f-8f928feeab8d\") " pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.920649 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.921954 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.923920 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.931628 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.951688 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-rnkvw"] Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.953826 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.974861 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.975239 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.975253 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 20:24:35 crc kubenswrapper[4726]: I1123 20:24:35.975318 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5dr84" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.003447 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rnkvw"] Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.011379 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-769a-account-create-qj28b"] Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.026163 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-combined-ca-bundle\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.026208 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pvfp\" (UniqueName: \"kubernetes.io/projected/51cac0d0-29b4-4867-9279-2142c34e613c-kube-api-access-5pvfp\") pod \"neutron-769a-account-create-qj28b\" (UID: \"51cac0d0-29b4-4867-9279-2142c34e613c\") " pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.026258 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c555bb2-ca8a-412e-852f-8f928feeab8d-operator-scripts\") pod \"neutron-db-create-psjjc\" (UID: \"5c555bb2-ca8a-412e-852f-8f928feeab8d\") " pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.026344 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zcsz\" (UniqueName: \"kubernetes.io/projected/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-kube-api-access-5zcsz\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.026362 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57msx\" (UniqueName: \"kubernetes.io/projected/5c555bb2-ca8a-412e-852f-8f928feeab8d-kube-api-access-57msx\") pod \"neutron-db-create-psjjc\" (UID: \"5c555bb2-ca8a-412e-852f-8f928feeab8d\") " pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.026386 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-config-data\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.026404 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51cac0d0-29b4-4867-9279-2142c34e613c-operator-scripts\") pod \"neutron-769a-account-create-qj28b\" (UID: \"51cac0d0-29b4-4867-9279-2142c34e613c\") " pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.027161 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c555bb2-ca8a-412e-852f-8f928feeab8d-operator-scripts\") pod \"neutron-db-create-psjjc\" (UID: \"5c555bb2-ca8a-412e-852f-8f928feeab8d\") " pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.070756 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57msx\" (UniqueName: \"kubernetes.io/projected/5c555bb2-ca8a-412e-852f-8f928feeab8d-kube-api-access-57msx\") pod \"neutron-db-create-psjjc\" (UID: \"5c555bb2-ca8a-412e-852f-8f928feeab8d\") " pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.087353 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.128085 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zcsz\" (UniqueName: \"kubernetes.io/projected/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-kube-api-access-5zcsz\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.128139 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-config-data\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.128158 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51cac0d0-29b4-4867-9279-2142c34e613c-operator-scripts\") pod \"neutron-769a-account-create-qj28b\" (UID: \"51cac0d0-29b4-4867-9279-2142c34e613c\") " pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.128178 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-combined-ca-bundle\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.128194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pvfp\" (UniqueName: \"kubernetes.io/projected/51cac0d0-29b4-4867-9279-2142c34e613c-kube-api-access-5pvfp\") pod \"neutron-769a-account-create-qj28b\" (UID: \"51cac0d0-29b4-4867-9279-2142c34e613c\") " pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.132577 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51cac0d0-29b4-4867-9279-2142c34e613c-operator-scripts\") pod \"neutron-769a-account-create-qj28b\" (UID: \"51cac0d0-29b4-4867-9279-2142c34e613c\") " pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.134676 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-config-data\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.134846 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-combined-ca-bundle\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.152509 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pvfp\" (UniqueName: \"kubernetes.io/projected/51cac0d0-29b4-4867-9279-2142c34e613c-kube-api-access-5pvfp\") pod \"neutron-769a-account-create-qj28b\" (UID: \"51cac0d0-29b4-4867-9279-2142c34e613c\") " pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.153534 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zcsz\" (UniqueName: \"kubernetes.io/projected/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-kube-api-access-5zcsz\") pod \"keystone-db-sync-rnkvw\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.232998 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-m4fr4"] Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.257740 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:36 crc kubenswrapper[4726]: W1123 20:24:36.294532 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b7c9ec1_e0d0_48cb_bfec_510f28a2a234.slice/crio-8ef71de5ae3a241b6ae13a6bb68b89503daccf24f59d0449a943eff3a9094629 WatchSource:0}: Error finding container 8ef71de5ae3a241b6ae13a6bb68b89503daccf24f59d0449a943eff3a9094629: Status 404 returned error can't find the container with id 8ef71de5ae3a241b6ae13a6bb68b89503daccf24f59d0449a943eff3a9094629 Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.309535 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.368740 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5d0d-account-create-tn2pk"] Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.698423 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-x7d9j"] Nov 23 20:24:36 crc kubenswrapper[4726]: I1123 20:24:36.730775 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b6a0-account-create-2zr8p"] Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:36.818877 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-psjjc"] Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:36.856140 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rnkvw"] Nov 23 20:24:37 crc kubenswrapper[4726]: W1123 20:24:36.882133 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d16b97d_7a0e_4312_94c8_f451d9e66d2b.slice/crio-ba27a1bb483da5ddff7e7d6155cedb53f7ffd0a71cef8a0e0631f52b1a68fbdd WatchSource:0}: Error finding container ba27a1bb483da5ddff7e7d6155cedb53f7ffd0a71cef8a0e0631f52b1a68fbdd: Status 404 returned error can't find the container with id ba27a1bb483da5ddff7e7d6155cedb53f7ffd0a71cef8a0e0631f52b1a68fbdd Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:36.921383 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-769a-account-create-qj28b"] Nov 23 20:24:37 crc kubenswrapper[4726]: W1123 20:24:36.938486 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51cac0d0_29b4_4867_9279_2142c34e613c.slice/crio-411229bdabbfb1fb32ac489e226116ddad0ca5ec7abc79558c0cb2cedf388964 WatchSource:0}: Error finding container 411229bdabbfb1fb32ac489e226116ddad0ca5ec7abc79558c0cb2cedf388964: Status 404 returned error can't find the container with id 411229bdabbfb1fb32ac489e226116ddad0ca5ec7abc79558c0cb2cedf388964 Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.008395 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.094379 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-ggvz7"] Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.095748 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" podUID="b34343ff-43aa-4518-a7d5-96c3b3678e18" containerName="dnsmasq-dns" containerID="cri-o://c66c311edcc2cbb5a34ff94d0ba78eb88b732a23d5e098970d620185822c64f0" gracePeriod=10 Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.292615 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rnkvw" event={"ID":"2d16b97d-7a0e-4312-94c8-f451d9e66d2b","Type":"ContainerStarted","Data":"ba27a1bb483da5ddff7e7d6155cedb53f7ffd0a71cef8a0e0631f52b1a68fbdd"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.322750 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b6a0-account-create-2zr8p" event={"ID":"1c555255-c599-4b91-9a76-c860a2c99228","Type":"ContainerStarted","Data":"13d56dbd9d7aa9fc206ba2b30c26eadececfea18f4c59621c62b00bf1dcd213f"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.322794 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b6a0-account-create-2zr8p" event={"ID":"1c555255-c599-4b91-9a76-c860a2c99228","Type":"ContainerStarted","Data":"d0b1cd782e455e5af8e2fb517735408dd0da64c30725c50ef438d9bfba716140"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.337472 4726 generic.go:334] "Generic (PLEG): container finished" podID="7b7c9ec1-e0d0-48cb-bfec-510f28a2a234" containerID="bd3dfe55a7358e627b47808f2d18e797ff6107b0a0ba799755b44a27b2c8600f" exitCode=0 Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.337540 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m4fr4" event={"ID":"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234","Type":"ContainerDied","Data":"bd3dfe55a7358e627b47808f2d18e797ff6107b0a0ba799755b44a27b2c8600f"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.337565 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m4fr4" event={"ID":"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234","Type":"ContainerStarted","Data":"8ef71de5ae3a241b6ae13a6bb68b89503daccf24f59d0449a943eff3a9094629"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.376677 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-b6a0-account-create-2zr8p" podStartSLOduration=2.376649531 podStartE2EDuration="2.376649531s" podCreationTimestamp="2025-11-23 20:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:24:37.366763997 +0000 UTC m=+985.515804963" watchObservedRunningTime="2025-11-23 20:24:37.376649531 +0000 UTC m=+985.525690487" Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.400442 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-psjjc" event={"ID":"5c555bb2-ca8a-412e-852f-8f928feeab8d","Type":"ContainerStarted","Data":"019216e2312cb4f4fb42b0f405f65053f655bca0deaddf59b331b79a020364fb"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.400491 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-psjjc" event={"ID":"5c555bb2-ca8a-412e-852f-8f928feeab8d","Type":"ContainerStarted","Data":"61a0800c04d10e4a00e704addd9dbb077af6625b4926d442ef15ec11740308bb"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.426451 4726 generic.go:334] "Generic (PLEG): container finished" podID="146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd" containerID="7a6b3d215f9277e5a0e3dd2b210271a307ecdc0c5926dfc2f7c4f324af667aad" exitCode=0 Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.426746 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5d0d-account-create-tn2pk" event={"ID":"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd","Type":"ContainerDied","Data":"7a6b3d215f9277e5a0e3dd2b210271a307ecdc0c5926dfc2f7c4f324af667aad"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.426773 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5d0d-account-create-tn2pk" event={"ID":"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd","Type":"ContainerStarted","Data":"8876273264ef6235d6225ea01e956d0e1f8751fd95c33fa1e8decb8419008d63"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.437909 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-psjjc" podStartSLOduration=2.437890837 podStartE2EDuration="2.437890837s" podCreationTimestamp="2025-11-23 20:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:24:37.426277055 +0000 UTC m=+985.575318011" watchObservedRunningTime="2025-11-23 20:24:37.437890837 +0000 UTC m=+985.586931803" Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.443993 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x7d9j" event={"ID":"aa225da7-111f-449f-8e6e-d2a73bb39a8f","Type":"ContainerStarted","Data":"b0f11e8a5604fa2461a74c448c4711802b50b13f152ad1cec2f187d7ced542d6"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.444029 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x7d9j" event={"ID":"aa225da7-111f-449f-8e6e-d2a73bb39a8f","Type":"ContainerStarted","Data":"0a328b4316647267f5226319c95a903392ff8b5c47d872246bddbd6274521890"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.478187 4726 generic.go:334] "Generic (PLEG): container finished" podID="b34343ff-43aa-4518-a7d5-96c3b3678e18" containerID="c66c311edcc2cbb5a34ff94d0ba78eb88b732a23d5e098970d620185822c64f0" exitCode=0 Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.478270 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" event={"ID":"b34343ff-43aa-4518-a7d5-96c3b3678e18","Type":"ContainerDied","Data":"c66c311edcc2cbb5a34ff94d0ba78eb88b732a23d5e098970d620185822c64f0"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.487787 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769a-account-create-qj28b" event={"ID":"51cac0d0-29b4-4867-9279-2142c34e613c","Type":"ContainerStarted","Data":"a44955defed0c124ad62cd79470d7c1fb78def6a01bfebecf3c07c0bde9d2bdd"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.487823 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769a-account-create-qj28b" event={"ID":"51cac0d0-29b4-4867-9279-2142c34e613c","Type":"ContainerStarted","Data":"411229bdabbfb1fb32ac489e226116ddad0ca5ec7abc79558c0cb2cedf388964"} Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.507202 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-x7d9j" podStartSLOduration=2.507179885 podStartE2EDuration="2.507179885s" podCreationTimestamp="2025-11-23 20:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:24:37.4860616 +0000 UTC m=+985.635102556" watchObservedRunningTime="2025-11-23 20:24:37.507179885 +0000 UTC m=+985.656220841" Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.540961 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-769a-account-create-qj28b" podStartSLOduration=2.54094136 podStartE2EDuration="2.54094136s" podCreationTimestamp="2025-11-23 20:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:24:37.521182493 +0000 UTC m=+985.670223449" watchObservedRunningTime="2025-11-23 20:24:37.54094136 +0000 UTC m=+985.689982326" Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.900947 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.985901 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-dns-svc\") pod \"b34343ff-43aa-4518-a7d5-96c3b3678e18\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.986365 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-nb\") pod \"b34343ff-43aa-4518-a7d5-96c3b3678e18\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.986409 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v86fd\" (UniqueName: \"kubernetes.io/projected/b34343ff-43aa-4518-a7d5-96c3b3678e18-kube-api-access-v86fd\") pod \"b34343ff-43aa-4518-a7d5-96c3b3678e18\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.987080 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-sb\") pod \"b34343ff-43aa-4518-a7d5-96c3b3678e18\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " Nov 23 20:24:37 crc kubenswrapper[4726]: I1123 20:24:37.987138 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-config\") pod \"b34343ff-43aa-4518-a7d5-96c3b3678e18\" (UID: \"b34343ff-43aa-4518-a7d5-96c3b3678e18\") " Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.005584 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b34343ff-43aa-4518-a7d5-96c3b3678e18-kube-api-access-v86fd" (OuterVolumeSpecName: "kube-api-access-v86fd") pod "b34343ff-43aa-4518-a7d5-96c3b3678e18" (UID: "b34343ff-43aa-4518-a7d5-96c3b3678e18"). InnerVolumeSpecName "kube-api-access-v86fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.053700 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b34343ff-43aa-4518-a7d5-96c3b3678e18" (UID: "b34343ff-43aa-4518-a7d5-96c3b3678e18"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.053739 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b34343ff-43aa-4518-a7d5-96c3b3678e18" (UID: "b34343ff-43aa-4518-a7d5-96c3b3678e18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.057317 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b34343ff-43aa-4518-a7d5-96c3b3678e18" (UID: "b34343ff-43aa-4518-a7d5-96c3b3678e18"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.092260 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.094404 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.094720 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.094778 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v86fd\" (UniqueName: \"kubernetes.io/projected/b34343ff-43aa-4518-a7d5-96c3b3678e18-kube-api-access-v86fd\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.107816 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-config" (OuterVolumeSpecName: "config") pod "b34343ff-43aa-4518-a7d5-96c3b3678e18" (UID: "b34343ff-43aa-4518-a7d5-96c3b3678e18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.199313 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b34343ff-43aa-4518-a7d5-96c3b3678e18-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.497268 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.497262 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-ggvz7" event={"ID":"b34343ff-43aa-4518-a7d5-96c3b3678e18","Type":"ContainerDied","Data":"417f9d4ebc33f7b0a7392dd7c3e45f26a9a184437e2be6500d2017fd44303225"} Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.497455 4726 scope.go:117] "RemoveContainer" containerID="c66c311edcc2cbb5a34ff94d0ba78eb88b732a23d5e098970d620185822c64f0" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.515068 4726 generic.go:334] "Generic (PLEG): container finished" podID="51cac0d0-29b4-4867-9279-2142c34e613c" containerID="a44955defed0c124ad62cd79470d7c1fb78def6a01bfebecf3c07c0bde9d2bdd" exitCode=0 Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.515171 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769a-account-create-qj28b" event={"ID":"51cac0d0-29b4-4867-9279-2142c34e613c","Type":"ContainerDied","Data":"a44955defed0c124ad62cd79470d7c1fb78def6a01bfebecf3c07c0bde9d2bdd"} Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.519144 4726 generic.go:334] "Generic (PLEG): container finished" podID="1c555255-c599-4b91-9a76-c860a2c99228" containerID="13d56dbd9d7aa9fc206ba2b30c26eadececfea18f4c59621c62b00bf1dcd213f" exitCode=0 Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.519277 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b6a0-account-create-2zr8p" event={"ID":"1c555255-c599-4b91-9a76-c860a2c99228","Type":"ContainerDied","Data":"13d56dbd9d7aa9fc206ba2b30c26eadececfea18f4c59621c62b00bf1dcd213f"} Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.521230 4726 generic.go:334] "Generic (PLEG): container finished" podID="5c555bb2-ca8a-412e-852f-8f928feeab8d" containerID="019216e2312cb4f4fb42b0f405f65053f655bca0deaddf59b331b79a020364fb" exitCode=0 Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.521283 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-psjjc" event={"ID":"5c555bb2-ca8a-412e-852f-8f928feeab8d","Type":"ContainerDied","Data":"019216e2312cb4f4fb42b0f405f65053f655bca0deaddf59b331b79a020364fb"} Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.533588 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-ggvz7"] Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.534784 4726 scope.go:117] "RemoveContainer" containerID="41f66e41cadc1443f27db47b7b26814ec13f82f321a96c7e2d7939b05e435e3f" Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.534987 4726 generic.go:334] "Generic (PLEG): container finished" podID="aa225da7-111f-449f-8e6e-d2a73bb39a8f" containerID="b0f11e8a5604fa2461a74c448c4711802b50b13f152ad1cec2f187d7ced542d6" exitCode=0 Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.535258 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x7d9j" event={"ID":"aa225da7-111f-449f-8e6e-d2a73bb39a8f","Type":"ContainerDied","Data":"b0f11e8a5604fa2461a74c448c4711802b50b13f152ad1cec2f187d7ced542d6"} Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.550292 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-ggvz7"] Nov 23 20:24:38 crc kubenswrapper[4726]: I1123 20:24:38.600723 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b34343ff-43aa-4518-a7d5-96c3b3678e18" path="/var/lib/kubelet/pods/b34343ff-43aa-4518-a7d5-96c3b3678e18/volumes" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.000071 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.068588 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.117218 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbj5s\" (UniqueName: \"kubernetes.io/projected/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-kube-api-access-cbj5s\") pod \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\" (UID: \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\") " Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.117335 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-operator-scripts\") pod \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\" (UID: \"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd\") " Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.118078 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd" (UID: "146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.123068 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-kube-api-access-cbj5s" (OuterVolumeSpecName: "kube-api-access-cbj5s") pod "146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd" (UID: "146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd"). InnerVolumeSpecName "kube-api-access-cbj5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.218701 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5sph\" (UniqueName: \"kubernetes.io/projected/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-kube-api-access-n5sph\") pod \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\" (UID: \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\") " Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.218808 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-operator-scripts\") pod \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\" (UID: \"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234\") " Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.219124 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.219142 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbj5s\" (UniqueName: \"kubernetes.io/projected/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd-kube-api-access-cbj5s\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.219503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7b7c9ec1-e0d0-48cb-bfec-510f28a2a234" (UID: "7b7c9ec1-e0d0-48cb-bfec-510f28a2a234"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.224543 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-kube-api-access-n5sph" (OuterVolumeSpecName: "kube-api-access-n5sph") pod "7b7c9ec1-e0d0-48cb-bfec-510f28a2a234" (UID: "7b7c9ec1-e0d0-48cb-bfec-510f28a2a234"). InnerVolumeSpecName "kube-api-access-n5sph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.320472 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5sph\" (UniqueName: \"kubernetes.io/projected/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-kube-api-access-n5sph\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.320512 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.553204 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m4fr4" event={"ID":"7b7c9ec1-e0d0-48cb-bfec-510f28a2a234","Type":"ContainerDied","Data":"8ef71de5ae3a241b6ae13a6bb68b89503daccf24f59d0449a943eff3a9094629"} Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.553252 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ef71de5ae3a241b6ae13a6bb68b89503daccf24f59d0449a943eff3a9094629" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.553313 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m4fr4" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.555908 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5d0d-account-create-tn2pk" Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.555860 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5d0d-account-create-tn2pk" event={"ID":"146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd","Type":"ContainerDied","Data":"8876273264ef6235d6225ea01e956d0e1f8751fd95c33fa1e8decb8419008d63"} Nov 23 20:24:39 crc kubenswrapper[4726]: I1123 20:24:39.556002 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8876273264ef6235d6225ea01e956d0e1f8751fd95c33fa1e8decb8419008d63" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.844602 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.869220 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.890769 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.903209 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.979901 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c555bb2-ca8a-412e-852f-8f928feeab8d-operator-scripts\") pod \"5c555bb2-ca8a-412e-852f-8f928feeab8d\" (UID: \"5c555bb2-ca8a-412e-852f-8f928feeab8d\") " Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.979970 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57msx\" (UniqueName: \"kubernetes.io/projected/5c555bb2-ca8a-412e-852f-8f928feeab8d-kube-api-access-57msx\") pod \"5c555bb2-ca8a-412e-852f-8f928feeab8d\" (UID: \"5c555bb2-ca8a-412e-852f-8f928feeab8d\") " Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.980005 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa225da7-111f-449f-8e6e-d2a73bb39a8f-operator-scripts\") pod \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\" (UID: \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\") " Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.980044 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pvfp\" (UniqueName: \"kubernetes.io/projected/51cac0d0-29b4-4867-9279-2142c34e613c-kube-api-access-5pvfp\") pod \"51cac0d0-29b4-4867-9279-2142c34e613c\" (UID: \"51cac0d0-29b4-4867-9279-2142c34e613c\") " Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.980088 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ddv4\" (UniqueName: \"kubernetes.io/projected/1c555255-c599-4b91-9a76-c860a2c99228-kube-api-access-7ddv4\") pod \"1c555255-c599-4b91-9a76-c860a2c99228\" (UID: \"1c555255-c599-4b91-9a76-c860a2c99228\") " Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.980182 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51cac0d0-29b4-4867-9279-2142c34e613c-operator-scripts\") pod \"51cac0d0-29b4-4867-9279-2142c34e613c\" (UID: \"51cac0d0-29b4-4867-9279-2142c34e613c\") " Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.980222 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c555255-c599-4b91-9a76-c860a2c99228-operator-scripts\") pod \"1c555255-c599-4b91-9a76-c860a2c99228\" (UID: \"1c555255-c599-4b91-9a76-c860a2c99228\") " Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.980271 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c2nl\" (UniqueName: \"kubernetes.io/projected/aa225da7-111f-449f-8e6e-d2a73bb39a8f-kube-api-access-7c2nl\") pod \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\" (UID: \"aa225da7-111f-449f-8e6e-d2a73bb39a8f\") " Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.980938 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa225da7-111f-449f-8e6e-d2a73bb39a8f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aa225da7-111f-449f-8e6e-d2a73bb39a8f" (UID: "aa225da7-111f-449f-8e6e-d2a73bb39a8f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.981288 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51cac0d0-29b4-4867-9279-2142c34e613c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51cac0d0-29b4-4867-9279-2142c34e613c" (UID: "51cac0d0-29b4-4867-9279-2142c34e613c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.981337 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c555255-c599-4b91-9a76-c860a2c99228-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c555255-c599-4b91-9a76-c860a2c99228" (UID: "1c555255-c599-4b91-9a76-c860a2c99228"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.981466 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c555bb2-ca8a-412e-852f-8f928feeab8d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c555bb2-ca8a-412e-852f-8f928feeab8d" (UID: "5c555bb2-ca8a-412e-852f-8f928feeab8d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.984658 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51cac0d0-29b4-4867-9279-2142c34e613c-kube-api-access-5pvfp" (OuterVolumeSpecName: "kube-api-access-5pvfp") pod "51cac0d0-29b4-4867-9279-2142c34e613c" (UID: "51cac0d0-29b4-4867-9279-2142c34e613c"). InnerVolumeSpecName "kube-api-access-5pvfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.985126 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c555255-c599-4b91-9a76-c860a2c99228-kube-api-access-7ddv4" (OuterVolumeSpecName: "kube-api-access-7ddv4") pod "1c555255-c599-4b91-9a76-c860a2c99228" (UID: "1c555255-c599-4b91-9a76-c860a2c99228"). InnerVolumeSpecName "kube-api-access-7ddv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.985247 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c555bb2-ca8a-412e-852f-8f928feeab8d-kube-api-access-57msx" (OuterVolumeSpecName: "kube-api-access-57msx") pod "5c555bb2-ca8a-412e-852f-8f928feeab8d" (UID: "5c555bb2-ca8a-412e-852f-8f928feeab8d"). InnerVolumeSpecName "kube-api-access-57msx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:42 crc kubenswrapper[4726]: I1123 20:24:42.986281 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa225da7-111f-449f-8e6e-d2a73bb39a8f-kube-api-access-7c2nl" (OuterVolumeSpecName: "kube-api-access-7c2nl") pod "aa225da7-111f-449f-8e6e-d2a73bb39a8f" (UID: "aa225da7-111f-449f-8e6e-d2a73bb39a8f"). InnerVolumeSpecName "kube-api-access-7c2nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.081925 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c555255-c599-4b91-9a76-c860a2c99228-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.081981 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c2nl\" (UniqueName: \"kubernetes.io/projected/aa225da7-111f-449f-8e6e-d2a73bb39a8f-kube-api-access-7c2nl\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.082001 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c555bb2-ca8a-412e-852f-8f928feeab8d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.082013 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57msx\" (UniqueName: \"kubernetes.io/projected/5c555bb2-ca8a-412e-852f-8f928feeab8d-kube-api-access-57msx\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.082025 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa225da7-111f-449f-8e6e-d2a73bb39a8f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.082061 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pvfp\" (UniqueName: \"kubernetes.io/projected/51cac0d0-29b4-4867-9279-2142c34e613c-kube-api-access-5pvfp\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.082072 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ddv4\" (UniqueName: \"kubernetes.io/projected/1c555255-c599-4b91-9a76-c860a2c99228-kube-api-access-7ddv4\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.082084 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51cac0d0-29b4-4867-9279-2142c34e613c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.592233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b6a0-account-create-2zr8p" event={"ID":"1c555255-c599-4b91-9a76-c860a2c99228","Type":"ContainerDied","Data":"d0b1cd782e455e5af8e2fb517735408dd0da64c30725c50ef438d9bfba716140"} Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.592298 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0b1cd782e455e5af8e2fb517735408dd0da64c30725c50ef438d9bfba716140" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.592259 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b6a0-account-create-2zr8p" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.594066 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-psjjc" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.594119 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-psjjc" event={"ID":"5c555bb2-ca8a-412e-852f-8f928feeab8d","Type":"ContainerDied","Data":"61a0800c04d10e4a00e704addd9dbb077af6625b4926d442ef15ec11740308bb"} Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.594171 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61a0800c04d10e4a00e704addd9dbb077af6625b4926d442ef15ec11740308bb" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.596145 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x7d9j" event={"ID":"aa225da7-111f-449f-8e6e-d2a73bb39a8f","Type":"ContainerDied","Data":"0a328b4316647267f5226319c95a903392ff8b5c47d872246bddbd6274521890"} Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.596188 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a328b4316647267f5226319c95a903392ff8b5c47d872246bddbd6274521890" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.596264 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x7d9j" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.614575 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rnkvw" event={"ID":"2d16b97d-7a0e-4312-94c8-f451d9e66d2b","Type":"ContainerStarted","Data":"c57a772b2139f336b20a1afbc000cb7a34e5bbd0db983bc5f3c46f84f843e7bc"} Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.623537 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-769a-account-create-qj28b" event={"ID":"51cac0d0-29b4-4867-9279-2142c34e613c","Type":"ContainerDied","Data":"411229bdabbfb1fb32ac489e226116ddad0ca5ec7abc79558c0cb2cedf388964"} Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.623573 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="411229bdabbfb1fb32ac489e226116ddad0ca5ec7abc79558c0cb2cedf388964" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.623631 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-769a-account-create-qj28b" Nov 23 20:24:43 crc kubenswrapper[4726]: I1123 20:24:43.639075 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-rnkvw" podStartSLOduration=2.79915886 podStartE2EDuration="8.639058765s" podCreationTimestamp="2025-11-23 20:24:35 +0000 UTC" firstStartedPulling="2025-11-23 20:24:36.897322257 +0000 UTC m=+985.046363213" lastFinishedPulling="2025-11-23 20:24:42.737222162 +0000 UTC m=+990.886263118" observedRunningTime="2025-11-23 20:24:43.634511809 +0000 UTC m=+991.783552765" watchObservedRunningTime="2025-11-23 20:24:43.639058765 +0000 UTC m=+991.788099721" Nov 23 20:24:47 crc kubenswrapper[4726]: I1123 20:24:47.658211 4726 generic.go:334] "Generic (PLEG): container finished" podID="2d16b97d-7a0e-4312-94c8-f451d9e66d2b" containerID="c57a772b2139f336b20a1afbc000cb7a34e5bbd0db983bc5f3c46f84f843e7bc" exitCode=0 Nov 23 20:24:47 crc kubenswrapper[4726]: I1123 20:24:47.658297 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rnkvw" event={"ID":"2d16b97d-7a0e-4312-94c8-f451d9e66d2b","Type":"ContainerDied","Data":"c57a772b2139f336b20a1afbc000cb7a34e5bbd0db983bc5f3c46f84f843e7bc"} Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.055324 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.188596 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-config-data\") pod \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.188836 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-combined-ca-bundle\") pod \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.190017 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zcsz\" (UniqueName: \"kubernetes.io/projected/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-kube-api-access-5zcsz\") pod \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\" (UID: \"2d16b97d-7a0e-4312-94c8-f451d9e66d2b\") " Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.208052 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-kube-api-access-5zcsz" (OuterVolumeSpecName: "kube-api-access-5zcsz") pod "2d16b97d-7a0e-4312-94c8-f451d9e66d2b" (UID: "2d16b97d-7a0e-4312-94c8-f451d9e66d2b"). InnerVolumeSpecName "kube-api-access-5zcsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.212018 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d16b97d-7a0e-4312-94c8-f451d9e66d2b" (UID: "2d16b97d-7a0e-4312-94c8-f451d9e66d2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.246017 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-config-data" (OuterVolumeSpecName: "config-data") pod "2d16b97d-7a0e-4312-94c8-f451d9e66d2b" (UID: "2d16b97d-7a0e-4312-94c8-f451d9e66d2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.291789 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.291830 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zcsz\" (UniqueName: \"kubernetes.io/projected/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-kube-api-access-5zcsz\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.291843 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d16b97d-7a0e-4312-94c8-f451d9e66d2b-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.684946 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rnkvw" event={"ID":"2d16b97d-7a0e-4312-94c8-f451d9e66d2b","Type":"ContainerDied","Data":"ba27a1bb483da5ddff7e7d6155cedb53f7ffd0a71cef8a0e0631f52b1a68fbdd"} Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.684981 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rnkvw" Nov 23 20:24:49 crc kubenswrapper[4726]: I1123 20:24:49.684982 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba27a1bb483da5ddff7e7d6155cedb53f7ffd0a71cef8a0e0631f52b1a68fbdd" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.037630 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-zjx6w"] Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038274 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa225da7-111f-449f-8e6e-d2a73bb39a8f" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038292 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa225da7-111f-449f-8e6e-d2a73bb39a8f" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038310 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51cac0d0-29b4-4867-9279-2142c34e613c" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038317 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="51cac0d0-29b4-4867-9279-2142c34e613c" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038328 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b34343ff-43aa-4518-a7d5-96c3b3678e18" containerName="init" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038334 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b34343ff-43aa-4518-a7d5-96c3b3678e18" containerName="init" Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038344 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d16b97d-7a0e-4312-94c8-f451d9e66d2b" containerName="keystone-db-sync" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038350 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d16b97d-7a0e-4312-94c8-f451d9e66d2b" containerName="keystone-db-sync" Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038359 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c555bb2-ca8a-412e-852f-8f928feeab8d" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038366 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c555bb2-ca8a-412e-852f-8f928feeab8d" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038377 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b34343ff-43aa-4518-a7d5-96c3b3678e18" containerName="dnsmasq-dns" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038383 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b34343ff-43aa-4518-a7d5-96c3b3678e18" containerName="dnsmasq-dns" Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038390 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b7c9ec1-e0d0-48cb-bfec-510f28a2a234" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038396 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b7c9ec1-e0d0-48cb-bfec-510f28a2a234" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038402 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038408 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: E1123 20:24:50.038434 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c555255-c599-4b91-9a76-c860a2c99228" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038440 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c555255-c599-4b91-9a76-c860a2c99228" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038596 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b34343ff-43aa-4518-a7d5-96c3b3678e18" containerName="dnsmasq-dns" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038608 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d16b97d-7a0e-4312-94c8-f451d9e66d2b" containerName="keystone-db-sync" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038618 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c555255-c599-4b91-9a76-c860a2c99228" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038627 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa225da7-111f-449f-8e6e-d2a73bb39a8f" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038637 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="51cac0d0-29b4-4867-9279-2142c34e613c" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038647 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd" containerName="mariadb-account-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038654 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b7c9ec1-e0d0-48cb-bfec-510f28a2a234" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.038664 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c555bb2-ca8a-412e-852f-8f928feeab8d" containerName="mariadb-database-create" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.039211 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.048101 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-l9n7n"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.049498 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.061267 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.061535 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.061616 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.075670 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5dr84" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.075698 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.101562 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-l9n7n"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.124946 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zjx6w"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209575 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj7dn\" (UniqueName: \"kubernetes.io/projected/37b4a13f-25f0-47f0-a391-355f744803ca-kube-api-access-qj7dn\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209623 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-combined-ca-bundle\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209642 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrglg\" (UniqueName: \"kubernetes.io/projected/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-kube-api-access-qrglg\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209660 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-sb\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209685 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-fernet-keys\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209704 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-config\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209717 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-scripts\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209750 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-nb\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209790 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-config-data\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209838 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-dns-svc\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.209855 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-credential-keys\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.281059 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4tm6n"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.282339 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.287199 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hqg9v" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.287389 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.291255 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.300646 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4tm6n"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.311554 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-dns-svc\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.311795 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-credential-keys\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.311989 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj7dn\" (UniqueName: \"kubernetes.io/projected/37b4a13f-25f0-47f0-a391-355f744803ca-kube-api-access-qj7dn\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.312271 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-combined-ca-bundle\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.312333 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrglg\" (UniqueName: \"kubernetes.io/projected/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-kube-api-access-qrglg\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.312392 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-sb\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.312467 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-fernet-keys\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.312536 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-config\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.312595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-scripts\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.312669 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-nb\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.312750 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-config-data\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.315951 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-dns-svc\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.316038 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-config\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.316488 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-nb\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.318291 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-sb\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.320460 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-credential-keys\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.320637 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-config-data\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.322152 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-fernet-keys\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.325789 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-scripts\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.350461 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-combined-ca-bundle\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.361444 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj7dn\" (UniqueName: \"kubernetes.io/projected/37b4a13f-25f0-47f0-a391-355f744803ca-kube-api-access-qj7dn\") pod \"dnsmasq-dns-b76c757b7-l9n7n\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.372339 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.392414 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.394905 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.397802 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrglg\" (UniqueName: \"kubernetes.io/projected/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-kube-api-access-qrglg\") pod \"keystone-bootstrap-zjx6w\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.402255 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.402607 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.414264 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-config\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.414563 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-combined-ca-bundle\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.414662 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q4mr\" (UniqueName: \"kubernetes.io/projected/595377fa-b0c8-440a-a6f9-4bce1ce40352-kube-api-access-2q4mr\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.437435 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.516755 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-run-httpd\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.517659 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-log-httpd\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.517835 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-scripts\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.517975 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-config-data\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.518130 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.523016 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-config\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.523985 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.524130 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz6ht\" (UniqueName: \"kubernetes.io/projected/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-kube-api-access-rz6ht\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.522938 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-fq2c6"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.534236 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.527487 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-config\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.546227 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-fgrl4" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.546454 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.547984 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-combined-ca-bundle\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.548128 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q4mr\" (UniqueName: \"kubernetes.io/projected/595377fa-b0c8-440a-a6f9-4bce1ce40352-kube-api-access-2q4mr\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.548706 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-fq2c6"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.551341 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.557509 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-combined-ca-bundle\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.572531 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q4mr\" (UniqueName: \"kubernetes.io/projected/595377fa-b0c8-440a-a6f9-4bce1ce40352-kube-api-access-2q4mr\") pod \"neutron-db-sync-4tm6n\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.605017 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-s7rvr"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.643175 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.655046 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.681681 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.682050 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.706034 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-v2kht" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.682881 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-config-data\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731379 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-combined-ca-bundle\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731456 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-config-data\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731496 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731525 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-scripts\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731585 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tm4q\" (UniqueName: \"kubernetes.io/projected/97c02ff5-89f0-4834-ab15-1d226f632e17-kube-api-access-5tm4q\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731638 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz6ht\" (UniqueName: \"kubernetes.io/projected/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-kube-api-access-rz6ht\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731671 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-db-sync-config-data\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731715 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x95f6\" (UniqueName: \"kubernetes.io/projected/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-kube-api-access-x95f6\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731746 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-run-httpd\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-etc-machine-id\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731811 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-db-sync-config-data\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731831 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-combined-ca-bundle\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731852 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-log-httpd\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.731892 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-scripts\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.737240 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-scripts\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.737731 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-run-httpd\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.740240 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-log-httpd\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.743958 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.751027 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.765147 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-s7rvr"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.783460 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz6ht\" (UniqueName: \"kubernetes.io/projected/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-kube-api-access-rz6ht\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.809759 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-config-data\") pod \"ceilometer-0\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.811816 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834652 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-config-data\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834703 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-scripts\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834729 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tm4q\" (UniqueName: \"kubernetes.io/projected/97c02ff5-89f0-4834-ab15-1d226f632e17-kube-api-access-5tm4q\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834777 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-db-sync-config-data\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834812 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x95f6\" (UniqueName: \"kubernetes.io/projected/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-kube-api-access-x95f6\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834841 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-etc-machine-id\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834879 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-db-sync-config-data\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834900 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-combined-ca-bundle\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.834935 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-combined-ca-bundle\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.836332 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-etc-machine-id\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.840167 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-db-sync-config-data\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.841515 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-combined-ca-bundle\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.852669 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-l9n7n"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.858296 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-config-data\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.858592 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-combined-ca-bundle\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.860300 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-db-sync-config-data\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.862387 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-scripts\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.889358 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-skj9z"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.890891 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.914610 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tm4q\" (UniqueName: \"kubernetes.io/projected/97c02ff5-89f0-4834-ab15-1d226f632e17-kube-api-access-5tm4q\") pod \"barbican-db-sync-s7rvr\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.932178 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-skj9z"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.932235 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x95f6\" (UniqueName: \"kubernetes.io/projected/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-kube-api-access-x95f6\") pod \"cinder-db-sync-fq2c6\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.946885 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qk592"] Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.950676 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qk592" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.965480 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.965671 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.965784 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f72c9" Nov 23 20:24:50 crc kubenswrapper[4726]: I1123 20:24:50.991429 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qk592"] Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.021987 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040423 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3f7da5-afd6-44b1-b402-f3de8c39b518-logs\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040502 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-nb\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040527 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-scripts\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-config\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040599 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-combined-ca-bundle\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040651 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-config-data\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040684 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rjjj\" (UniqueName: \"kubernetes.io/projected/80352774-1f1a-4e54-b4e2-2fad013814e8-kube-api-access-2rjjj\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040832 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-dns-svc\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsd7d\" (UniqueName: \"kubernetes.io/projected/0c3f7da5-afd6-44b1-b402-f3de8c39b518-kube-api-access-rsd7d\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.040933 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-sb\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143463 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-dns-svc\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsd7d\" (UniqueName: \"kubernetes.io/projected/0c3f7da5-afd6-44b1-b402-f3de8c39b518-kube-api-access-rsd7d\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143543 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-sb\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143578 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3f7da5-afd6-44b1-b402-f3de8c39b518-logs\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143600 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-nb\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143617 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-scripts\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143632 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-config\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143653 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-combined-ca-bundle\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143675 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-config-data\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.143711 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rjjj\" (UniqueName: \"kubernetes.io/projected/80352774-1f1a-4e54-b4e2-2fad013814e8-kube-api-access-2rjjj\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.144295 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-dns-svc\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.144977 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-sb\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.144989 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-nb\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.145224 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3f7da5-afd6-44b1-b402-f3de8c39b518-logs\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.145791 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-config\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.162276 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-combined-ca-bundle\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.173319 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsd7d\" (UniqueName: \"kubernetes.io/projected/0c3f7da5-afd6-44b1-b402-f3de8c39b518-kube-api-access-rsd7d\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.174759 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rjjj\" (UniqueName: \"kubernetes.io/projected/80352774-1f1a-4e54-b4e2-2fad013814e8-kube-api-access-2rjjj\") pod \"dnsmasq-dns-66f4bdbdb7-skj9z\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.175275 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-config-data\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.179249 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-scripts\") pod \"placement-db-sync-qk592\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.193285 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.252714 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-l9n7n"] Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.267328 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.303018 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qk592" Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.476077 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.500400 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4tm6n"] Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.754261 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4tm6n" event={"ID":"595377fa-b0c8-440a-a6f9-4bce1ce40352","Type":"ContainerStarted","Data":"2bbcaa3e95542be9f390800249cb0388a83370bf11a1677f9dac3c529c440f61"} Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.762326 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" event={"ID":"37b4a13f-25f0-47f0-a391-355f744803ca","Type":"ContainerStarted","Data":"99c4162f0c11a4309d2b35705656ee079c114d276424bb91f542fc10c579034c"} Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.768510 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerStarted","Data":"436aec86f2a0e4ef14ed59d529e9dbcde35f31327aafce2a73eb2efd91f90456"} Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.769205 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-s7rvr"] Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.779888 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zjx6w"] Nov 23 20:24:51 crc kubenswrapper[4726]: I1123 20:24:51.994654 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-fq2c6"] Nov 23 20:24:52 crc kubenswrapper[4726]: W1123 20:24:52.010084 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3d6b028_7e7a_4c1d_a05e_5f49976b823d.slice/crio-678c2b73fdba7928193bbeee4dc96ac997490c6b90bc23b4696ea88f13ebf102 WatchSource:0}: Error finding container 678c2b73fdba7928193bbeee4dc96ac997490c6b90bc23b4696ea88f13ebf102: Status 404 returned error can't find the container with id 678c2b73fdba7928193bbeee4dc96ac997490c6b90bc23b4696ea88f13ebf102 Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.203625 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-skj9z"] Nov 23 20:24:52 crc kubenswrapper[4726]: W1123 20:24:52.222730 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80352774_1f1a_4e54_b4e2_2fad013814e8.slice/crio-d0471f4b6bcd416320fa8e86289fc560a80ce74ba68d8e4b35c5ae4b8ad09942 WatchSource:0}: Error finding container d0471f4b6bcd416320fa8e86289fc560a80ce74ba68d8e4b35c5ae4b8ad09942: Status 404 returned error can't find the container with id d0471f4b6bcd416320fa8e86289fc560a80ce74ba68d8e4b35c5ae4b8ad09942 Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.239073 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qk592"] Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.814371 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qk592" event={"ID":"0c3f7da5-afd6-44b1-b402-f3de8c39b518","Type":"ContainerStarted","Data":"5d984993c31f6e5024c00c370fc01757bb75ed5bf1649bef2025fa584796daf3"} Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.854564 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zjx6w" event={"ID":"c951a30f-9201-4e34-9fb2-9c0c2c9b206b","Type":"ContainerStarted","Data":"db4e8ff916ad6ec86fe213933be0253ba46268725cce9a2601df99d2f387c4fa"} Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.854615 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zjx6w" event={"ID":"c951a30f-9201-4e34-9fb2-9c0c2c9b206b","Type":"ContainerStarted","Data":"cca9b8c65aa7877284844c0de1ab6419ac8b6a36455686eb7340af6b1815f248"} Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.858769 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fq2c6" event={"ID":"b3d6b028-7e7a-4c1d-a05e-5f49976b823d","Type":"ContainerStarted","Data":"678c2b73fdba7928193bbeee4dc96ac997490c6b90bc23b4696ea88f13ebf102"} Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.863038 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s7rvr" event={"ID":"97c02ff5-89f0-4834-ab15-1d226f632e17","Type":"ContainerStarted","Data":"0c830b3579757cc5270044eba0092933d63b36e0784c2906e1fecfec26cd7fab"} Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.890681 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4tm6n" event={"ID":"595377fa-b0c8-440a-a6f9-4bce1ce40352","Type":"ContainerStarted","Data":"156ea67dd89b7b3cbe69b38612407c2015921a70c503bdf55fcbfc3a2e8c8960"} Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.911004 4726 generic.go:334] "Generic (PLEG): container finished" podID="80352774-1f1a-4e54-b4e2-2fad013814e8" containerID="7748370b657c9f474127749f38bba264a2fe20502c2cd6e0cb805cd13904285d" exitCode=0 Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.911094 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" event={"ID":"80352774-1f1a-4e54-b4e2-2fad013814e8","Type":"ContainerDied","Data":"7748370b657c9f474127749f38bba264a2fe20502c2cd6e0cb805cd13904285d"} Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.911118 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" event={"ID":"80352774-1f1a-4e54-b4e2-2fad013814e8","Type":"ContainerStarted","Data":"d0471f4b6bcd416320fa8e86289fc560a80ce74ba68d8e4b35c5ae4b8ad09942"} Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.941619 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4tm6n" podStartSLOduration=2.941605247 podStartE2EDuration="2.941605247s" podCreationTimestamp="2025-11-23 20:24:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:24:52.935697183 +0000 UTC m=+1001.084738139" watchObservedRunningTime="2025-11-23 20:24:52.941605247 +0000 UTC m=+1001.090646203" Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.942887 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-zjx6w" podStartSLOduration=3.942882022 podStartE2EDuration="3.942882022s" podCreationTimestamp="2025-11-23 20:24:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:24:52.910257009 +0000 UTC m=+1001.059297965" watchObservedRunningTime="2025-11-23 20:24:52.942882022 +0000 UTC m=+1001.091922978" Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.958022 4726 generic.go:334] "Generic (PLEG): container finished" podID="37b4a13f-25f0-47f0-a391-355f744803ca" containerID="0829c046daa61252c856422c6f39b64777d8fc93c0592333a65a46cd7356012e" exitCode=0 Nov 23 20:24:52 crc kubenswrapper[4726]: I1123 20:24:52.958063 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" event={"ID":"37b4a13f-25f0-47f0-a391-355f744803ca","Type":"ContainerDied","Data":"0829c046daa61252c856422c6f39b64777d8fc93c0592333a65a46cd7356012e"} Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.019880 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.537407 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.646808 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-nb\") pod \"37b4a13f-25f0-47f0-a391-355f744803ca\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.646894 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj7dn\" (UniqueName: \"kubernetes.io/projected/37b4a13f-25f0-47f0-a391-355f744803ca-kube-api-access-qj7dn\") pod \"37b4a13f-25f0-47f0-a391-355f744803ca\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.646961 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-config\") pod \"37b4a13f-25f0-47f0-a391-355f744803ca\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.647002 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-dns-svc\") pod \"37b4a13f-25f0-47f0-a391-355f744803ca\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.647120 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-sb\") pod \"37b4a13f-25f0-47f0-a391-355f744803ca\" (UID: \"37b4a13f-25f0-47f0-a391-355f744803ca\") " Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.672496 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37b4a13f-25f0-47f0-a391-355f744803ca" (UID: "37b4a13f-25f0-47f0-a391-355f744803ca"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.678599 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b4a13f-25f0-47f0-a391-355f744803ca-kube-api-access-qj7dn" (OuterVolumeSpecName: "kube-api-access-qj7dn") pod "37b4a13f-25f0-47f0-a391-355f744803ca" (UID: "37b4a13f-25f0-47f0-a391-355f744803ca"). InnerVolumeSpecName "kube-api-access-qj7dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.703299 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37b4a13f-25f0-47f0-a391-355f744803ca" (UID: "37b4a13f-25f0-47f0-a391-355f744803ca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.718470 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-config" (OuterVolumeSpecName: "config") pod "37b4a13f-25f0-47f0-a391-355f744803ca" (UID: "37b4a13f-25f0-47f0-a391-355f744803ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.745307 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37b4a13f-25f0-47f0-a391-355f744803ca" (UID: "37b4a13f-25f0-47f0-a391-355f744803ca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.750621 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.750737 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.750793 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj7dn\" (UniqueName: \"kubernetes.io/projected/37b4a13f-25f0-47f0-a391-355f744803ca-kube-api-access-qj7dn\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.750852 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.750932 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b4a13f-25f0-47f0-a391-355f744803ca-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.972124 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" event={"ID":"80352774-1f1a-4e54-b4e2-2fad013814e8","Type":"ContainerStarted","Data":"3f37211739e8493c4df510d6b3dba50a595edf0fefd4ac39c053103a173bf29d"} Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.972971 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.977407 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.979438 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b76c757b7-l9n7n" event={"ID":"37b4a13f-25f0-47f0-a391-355f744803ca","Type":"ContainerDied","Data":"99c4162f0c11a4309d2b35705656ee079c114d276424bb91f542fc10c579034c"} Nov 23 20:24:53 crc kubenswrapper[4726]: I1123 20:24:53.979485 4726 scope.go:117] "RemoveContainer" containerID="0829c046daa61252c856422c6f39b64777d8fc93c0592333a65a46cd7356012e" Nov 23 20:24:54 crc kubenswrapper[4726]: I1123 20:24:54.006784 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" podStartSLOduration=4.006756121 podStartE2EDuration="4.006756121s" podCreationTimestamp="2025-11-23 20:24:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:24:53.990734848 +0000 UTC m=+1002.139775814" watchObservedRunningTime="2025-11-23 20:24:54.006756121 +0000 UTC m=+1002.155797097" Nov 23 20:24:54 crc kubenswrapper[4726]: I1123 20:24:54.066447 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-l9n7n"] Nov 23 20:24:54 crc kubenswrapper[4726]: I1123 20:24:54.075826 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b76c757b7-l9n7n"] Nov 23 20:24:54 crc kubenswrapper[4726]: I1123 20:24:54.611966 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b4a13f-25f0-47f0-a391-355f744803ca" path="/var/lib/kubelet/pods/37b4a13f-25f0-47f0-a391-355f744803ca/volumes" Nov 23 20:24:59 crc kubenswrapper[4726]: I1123 20:24:59.024626 4726 generic.go:334] "Generic (PLEG): container finished" podID="c951a30f-9201-4e34-9fb2-9c0c2c9b206b" containerID="db4e8ff916ad6ec86fe213933be0253ba46268725cce9a2601df99d2f387c4fa" exitCode=0 Nov 23 20:24:59 crc kubenswrapper[4726]: I1123 20:24:59.024712 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zjx6w" event={"ID":"c951a30f-9201-4e34-9fb2-9c0c2c9b206b","Type":"ContainerDied","Data":"db4e8ff916ad6ec86fe213933be0253ba46268725cce9a2601df99d2f387c4fa"} Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.591808 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.702454 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-config-data\") pod \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.702761 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-combined-ca-bundle\") pod \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.702972 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-credential-keys\") pod \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.703085 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-fernet-keys\") pod \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.703230 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrglg\" (UniqueName: \"kubernetes.io/projected/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-kube-api-access-qrglg\") pod \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.703365 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-scripts\") pod \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\" (UID: \"c951a30f-9201-4e34-9fb2-9c0c2c9b206b\") " Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.708177 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c951a30f-9201-4e34-9fb2-9c0c2c9b206b" (UID: "c951a30f-9201-4e34-9fb2-9c0c2c9b206b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.721761 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-scripts" (OuterVolumeSpecName: "scripts") pod "c951a30f-9201-4e34-9fb2-9c0c2c9b206b" (UID: "c951a30f-9201-4e34-9fb2-9c0c2c9b206b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.721889 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c951a30f-9201-4e34-9fb2-9c0c2c9b206b" (UID: "c951a30f-9201-4e34-9fb2-9c0c2c9b206b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.725302 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-kube-api-access-qrglg" (OuterVolumeSpecName: "kube-api-access-qrglg") pod "c951a30f-9201-4e34-9fb2-9c0c2c9b206b" (UID: "c951a30f-9201-4e34-9fb2-9c0c2c9b206b"). InnerVolumeSpecName "kube-api-access-qrglg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.726810 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-config-data" (OuterVolumeSpecName: "config-data") pod "c951a30f-9201-4e34-9fb2-9c0c2c9b206b" (UID: "c951a30f-9201-4e34-9fb2-9c0c2c9b206b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.731646 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c951a30f-9201-4e34-9fb2-9c0c2c9b206b" (UID: "c951a30f-9201-4e34-9fb2-9c0c2c9b206b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.805651 4726 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.805686 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.805697 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrglg\" (UniqueName: \"kubernetes.io/projected/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-kube-api-access-qrglg\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.805707 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.805716 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:00 crc kubenswrapper[4726]: I1123 20:25:00.805723 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c951a30f-9201-4e34-9fb2-9c0c2c9b206b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.044164 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zjx6w" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.044186 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zjx6w" event={"ID":"c951a30f-9201-4e34-9fb2-9c0c2c9b206b","Type":"ContainerDied","Data":"cca9b8c65aa7877284844c0de1ab6419ac8b6a36455686eb7340af6b1815f248"} Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.044223 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cca9b8c65aa7877284844c0de1ab6419ac8b6a36455686eb7340af6b1815f248" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.115839 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-zjx6w"] Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.125065 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-zjx6w"] Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.214557 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-gcxk9"] Nov 23 20:25:01 crc kubenswrapper[4726]: E1123 20:25:01.215003 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b4a13f-25f0-47f0-a391-355f744803ca" containerName="init" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.215015 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b4a13f-25f0-47f0-a391-355f744803ca" containerName="init" Nov 23 20:25:01 crc kubenswrapper[4726]: E1123 20:25:01.215025 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c951a30f-9201-4e34-9fb2-9c0c2c9b206b" containerName="keystone-bootstrap" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.215033 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c951a30f-9201-4e34-9fb2-9c0c2c9b206b" containerName="keystone-bootstrap" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.215181 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c951a30f-9201-4e34-9fb2-9c0c2c9b206b" containerName="keystone-bootstrap" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.215202 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b4a13f-25f0-47f0-a391-355f744803ca" containerName="init" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.215759 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.218022 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.218448 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.218633 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.220058 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5dr84" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.223114 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gcxk9"] Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.223713 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.269132 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.325901 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-credential-keys\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.326039 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-scripts\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.326084 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-combined-ca-bundle\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.326153 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhg5r\" (UniqueName: \"kubernetes.io/projected/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-kube-api-access-zhg5r\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.326530 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-fernet-keys\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.326594 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-config-data\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.354502 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-569d458467-ffqcl"] Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.355293 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-569d458467-ffqcl" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="dnsmasq-dns" containerID="cri-o://e658d0749861ccc937b6325f5382fbd538c49a561a8aae44ab581690dd14187b" gracePeriod=10 Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.428916 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-credential-keys\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.428996 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-scripts\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.429027 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-combined-ca-bundle\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.429057 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhg5r\" (UniqueName: \"kubernetes.io/projected/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-kube-api-access-zhg5r\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.429091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-fernet-keys\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.429112 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-config-data\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.435518 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-config-data\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.452683 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-fernet-keys\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.453634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-scripts\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.453986 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-combined-ca-bundle\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.454523 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-credential-keys\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.462494 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhg5r\" (UniqueName: \"kubernetes.io/projected/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-kube-api-access-zhg5r\") pod \"keystone-bootstrap-gcxk9\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:01 crc kubenswrapper[4726]: I1123 20:25:01.529809 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:02 crc kubenswrapper[4726]: I1123 20:25:02.008198 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-569d458467-ffqcl" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Nov 23 20:25:02 crc kubenswrapper[4726]: I1123 20:25:02.058053 4726 generic.go:334] "Generic (PLEG): container finished" podID="862ffc28-0d10-4190-9654-51ae9590043e" containerID="e658d0749861ccc937b6325f5382fbd538c49a561a8aae44ab581690dd14187b" exitCode=0 Nov 23 20:25:02 crc kubenswrapper[4726]: I1123 20:25:02.058074 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-ffqcl" event={"ID":"862ffc28-0d10-4190-9654-51ae9590043e","Type":"ContainerDied","Data":"e658d0749861ccc937b6325f5382fbd538c49a561a8aae44ab581690dd14187b"} Nov 23 20:25:02 crc kubenswrapper[4726]: I1123 20:25:02.603124 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c951a30f-9201-4e34-9fb2-9c0c2c9b206b" path="/var/lib/kubelet/pods/c951a30f-9201-4e34-9fb2-9c0c2c9b206b/volumes" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.005588 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-569d458467-ffqcl" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Nov 23 20:25:12 crc kubenswrapper[4726]: E1123 20:25:12.096947 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645" Nov 23 20:25:12 crc kubenswrapper[4726]: E1123 20:25:12.097096 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5tm4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-s7rvr_openstack(97c02ff5-89f0-4834-ab15-1d226f632e17): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:25:12 crc kubenswrapper[4726]: E1123 20:25:12.098813 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-s7rvr" podUID="97c02ff5-89f0-4834-ab15-1d226f632e17" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.177724 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-569d458467-ffqcl" event={"ID":"862ffc28-0d10-4190-9654-51ae9590043e","Type":"ContainerDied","Data":"4b673c4a6931aa78f246965a6250664acfb63a23ceed1f96438f8b9625560eac"} Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.177805 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b673c4a6931aa78f246965a6250664acfb63a23ceed1f96438f8b9625560eac" Nov 23 20:25:12 crc kubenswrapper[4726]: E1123 20:25:12.179933 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645\\\"\"" pod="openstack/barbican-db-sync-s7rvr" podUID="97c02ff5-89f0-4834-ab15-1d226f632e17" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.197860 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.292906 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-dns-svc\") pod \"862ffc28-0d10-4190-9654-51ae9590043e\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.292968 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-nb\") pod \"862ffc28-0d10-4190-9654-51ae9590043e\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.292994 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbj86\" (UniqueName: \"kubernetes.io/projected/862ffc28-0d10-4190-9654-51ae9590043e-kube-api-access-bbj86\") pod \"862ffc28-0d10-4190-9654-51ae9590043e\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.293037 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-sb\") pod \"862ffc28-0d10-4190-9654-51ae9590043e\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.293214 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-config\") pod \"862ffc28-0d10-4190-9654-51ae9590043e\" (UID: \"862ffc28-0d10-4190-9654-51ae9590043e\") " Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.307247 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862ffc28-0d10-4190-9654-51ae9590043e-kube-api-access-bbj86" (OuterVolumeSpecName: "kube-api-access-bbj86") pod "862ffc28-0d10-4190-9654-51ae9590043e" (UID: "862ffc28-0d10-4190-9654-51ae9590043e"). InnerVolumeSpecName "kube-api-access-bbj86". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.348838 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "862ffc28-0d10-4190-9654-51ae9590043e" (UID: "862ffc28-0d10-4190-9654-51ae9590043e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.348847 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "862ffc28-0d10-4190-9654-51ae9590043e" (UID: "862ffc28-0d10-4190-9654-51ae9590043e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.352115 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-config" (OuterVolumeSpecName: "config") pod "862ffc28-0d10-4190-9654-51ae9590043e" (UID: "862ffc28-0d10-4190-9654-51ae9590043e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.367081 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "862ffc28-0d10-4190-9654-51ae9590043e" (UID: "862ffc28-0d10-4190-9654-51ae9590043e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.395232 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.395265 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.395278 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbj86\" (UniqueName: \"kubernetes.io/projected/862ffc28-0d10-4190-9654-51ae9590043e-kube-api-access-bbj86\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.395287 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:12 crc kubenswrapper[4726]: I1123 20:25:12.395297 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862ffc28-0d10-4190-9654-51ae9590043e-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:13 crc kubenswrapper[4726]: I1123 20:25:13.187950 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-569d458467-ffqcl" Nov 23 20:25:13 crc kubenswrapper[4726]: I1123 20:25:13.213120 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-569d458467-ffqcl"] Nov 23 20:25:13 crc kubenswrapper[4726]: I1123 20:25:13.221104 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-569d458467-ffqcl"] Nov 23 20:25:14 crc kubenswrapper[4726]: I1123 20:25:14.603538 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862ffc28-0d10-4190-9654-51ae9590043e" path="/var/lib/kubelet/pods/862ffc28-0d10-4190-9654-51ae9590043e/volumes" Nov 23 20:25:15 crc kubenswrapper[4726]: E1123 20:25:15.326812 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879" Nov 23 20:25:15 crc kubenswrapper[4726]: E1123 20:25:15.327565 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x95f6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-fq2c6_openstack(b3d6b028-7e7a-4c1d-a05e-5f49976b823d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 20:25:15 crc kubenswrapper[4726]: E1123 20:25:15.329507 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-fq2c6" podUID="b3d6b028-7e7a-4c1d-a05e-5f49976b823d" Nov 23 20:25:15 crc kubenswrapper[4726]: I1123 20:25:15.757352 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gcxk9"] Nov 23 20:25:15 crc kubenswrapper[4726]: I1123 20:25:15.767065 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 23 20:25:16 crc kubenswrapper[4726]: I1123 20:25:16.220478 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerStarted","Data":"27d51503001e6dc16b2018d4a060cc21f99e42ec9d6537683ff985b7654e9697"} Nov 23 20:25:16 crc kubenswrapper[4726]: I1123 20:25:16.222190 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gcxk9" event={"ID":"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4","Type":"ContainerStarted","Data":"d7a28f3138cbde6b4e68d7490eeef283ac368027bf675200afddf14efa2054ff"} Nov 23 20:25:16 crc kubenswrapper[4726]: I1123 20:25:16.222235 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gcxk9" event={"ID":"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4","Type":"ContainerStarted","Data":"fd08c571e7be6742f653f1d49da2b8eb5af3e46dbb011e29b95039489f364819"} Nov 23 20:25:16 crc kubenswrapper[4726]: I1123 20:25:16.236268 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qk592" event={"ID":"0c3f7da5-afd6-44b1-b402-f3de8c39b518","Type":"ContainerStarted","Data":"61c6c2fdc4e47dac3d875f5aaf8e00b45b2bb58a94ec35bed7f245267e41db62"} Nov 23 20:25:16 crc kubenswrapper[4726]: E1123 20:25:16.240649 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879\\\"\"" pod="openstack/cinder-db-sync-fq2c6" podUID="b3d6b028-7e7a-4c1d-a05e-5f49976b823d" Nov 23 20:25:16 crc kubenswrapper[4726]: I1123 20:25:16.247382 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-gcxk9" podStartSLOduration=15.247364884 podStartE2EDuration="15.247364884s" podCreationTimestamp="2025-11-23 20:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:16.237728837 +0000 UTC m=+1024.386769803" watchObservedRunningTime="2025-11-23 20:25:16.247364884 +0000 UTC m=+1024.396405840" Nov 23 20:25:16 crc kubenswrapper[4726]: I1123 20:25:16.273799 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qk592" podStartSLOduration=6.436081303 podStartE2EDuration="26.273783465s" podCreationTimestamp="2025-11-23 20:24:50 +0000 UTC" firstStartedPulling="2025-11-23 20:24:52.252937976 +0000 UTC m=+1000.401978932" lastFinishedPulling="2025-11-23 20:25:12.090640138 +0000 UTC m=+1020.239681094" observedRunningTime="2025-11-23 20:25:16.26600225 +0000 UTC m=+1024.415043236" watchObservedRunningTime="2025-11-23 20:25:16.273783465 +0000 UTC m=+1024.422824431" Nov 23 20:25:17 crc kubenswrapper[4726]: I1123 20:25:17.006788 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-569d458467-ffqcl" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Nov 23 20:25:17 crc kubenswrapper[4726]: I1123 20:25:17.250267 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerStarted","Data":"48f4363e2fa89c5a8408a6a7c21589d4c82008c2e40bf3f98d946ad9dc5b6304"} Nov 23 20:25:18 crc kubenswrapper[4726]: I1123 20:25:18.259377 4726 generic.go:334] "Generic (PLEG): container finished" podID="0c3f7da5-afd6-44b1-b402-f3de8c39b518" containerID="61c6c2fdc4e47dac3d875f5aaf8e00b45b2bb58a94ec35bed7f245267e41db62" exitCode=0 Nov 23 20:25:18 crc kubenswrapper[4726]: I1123 20:25:18.259558 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qk592" event={"ID":"0c3f7da5-afd6-44b1-b402-f3de8c39b518","Type":"ContainerDied","Data":"61c6c2fdc4e47dac3d875f5aaf8e00b45b2bb58a94ec35bed7f245267e41db62"} Nov 23 20:25:19 crc kubenswrapper[4726]: I1123 20:25:19.268002 4726 generic.go:334] "Generic (PLEG): container finished" podID="595377fa-b0c8-440a-a6f9-4bce1ce40352" containerID="156ea67dd89b7b3cbe69b38612407c2015921a70c503bdf55fcbfc3a2e8c8960" exitCode=0 Nov 23 20:25:19 crc kubenswrapper[4726]: I1123 20:25:19.268176 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4tm6n" event={"ID":"595377fa-b0c8-440a-a6f9-4bce1ce40352","Type":"ContainerDied","Data":"156ea67dd89b7b3cbe69b38612407c2015921a70c503bdf55fcbfc3a2e8c8960"} Nov 23 20:25:21 crc kubenswrapper[4726]: I1123 20:25:21.300082 4726 generic.go:334] "Generic (PLEG): container finished" podID="c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" containerID="d7a28f3138cbde6b4e68d7490eeef283ac368027bf675200afddf14efa2054ff" exitCode=0 Nov 23 20:25:21 crc kubenswrapper[4726]: I1123 20:25:21.300184 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gcxk9" event={"ID":"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4","Type":"ContainerDied","Data":"d7a28f3138cbde6b4e68d7490eeef283ac368027bf675200afddf14efa2054ff"} Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.730541 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qk592" Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.850917 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.863641 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.904923 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3f7da5-afd6-44b1-b402-f3de8c39b518-logs\") pod \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.904994 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-scripts\") pod \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.905358 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c3f7da5-afd6-44b1-b402-f3de8c39b518-logs" (OuterVolumeSpecName: "logs") pod "0c3f7da5-afd6-44b1-b402-f3de8c39b518" (UID: "0c3f7da5-afd6-44b1-b402-f3de8c39b518"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.906132 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-config-data\") pod \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.906200 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-combined-ca-bundle\") pod \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.906250 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsd7d\" (UniqueName: \"kubernetes.io/projected/0c3f7da5-afd6-44b1-b402-f3de8c39b518-kube-api-access-rsd7d\") pod \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\" (UID: \"0c3f7da5-afd6-44b1-b402-f3de8c39b518\") " Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.907018 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3f7da5-afd6-44b1-b402-f3de8c39b518-logs\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.911001 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c3f7da5-afd6-44b1-b402-f3de8c39b518-kube-api-access-rsd7d" (OuterVolumeSpecName: "kube-api-access-rsd7d") pod "0c3f7da5-afd6-44b1-b402-f3de8c39b518" (UID: "0c3f7da5-afd6-44b1-b402-f3de8c39b518"). InnerVolumeSpecName "kube-api-access-rsd7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.927718 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-scripts" (OuterVolumeSpecName: "scripts") pod "0c3f7da5-afd6-44b1-b402-f3de8c39b518" (UID: "0c3f7da5-afd6-44b1-b402-f3de8c39b518"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.943670 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c3f7da5-afd6-44b1-b402-f3de8c39b518" (UID: "0c3f7da5-afd6-44b1-b402-f3de8c39b518"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:22 crc kubenswrapper[4726]: I1123 20:25:22.954922 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-config-data" (OuterVolumeSpecName: "config-data") pod "0c3f7da5-afd6-44b1-b402-f3de8c39b518" (UID: "0c3f7da5-afd6-44b1-b402-f3de8c39b518"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.009229 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q4mr\" (UniqueName: \"kubernetes.io/projected/595377fa-b0c8-440a-a6f9-4bce1ce40352-kube-api-access-2q4mr\") pod \"595377fa-b0c8-440a-a6f9-4bce1ce40352\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.010009 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-combined-ca-bundle\") pod \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.010656 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-scripts\") pod \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.010807 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-config\") pod \"595377fa-b0c8-440a-a6f9-4bce1ce40352\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.010840 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-credential-keys\") pod \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.010877 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhg5r\" (UniqueName: \"kubernetes.io/projected/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-kube-api-access-zhg5r\") pod \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.010929 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-fernet-keys\") pod \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.010950 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-combined-ca-bundle\") pod \"595377fa-b0c8-440a-a6f9-4bce1ce40352\" (UID: \"595377fa-b0c8-440a-a6f9-4bce1ce40352\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.010988 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-config-data\") pod \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\" (UID: \"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4\") " Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.011542 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.011559 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.011568 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3f7da5-afd6-44b1-b402-f3de8c39b518-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.011597 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsd7d\" (UniqueName: \"kubernetes.io/projected/0c3f7da5-afd6-44b1-b402-f3de8c39b518-kube-api-access-rsd7d\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.013454 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/595377fa-b0c8-440a-a6f9-4bce1ce40352-kube-api-access-2q4mr" (OuterVolumeSpecName: "kube-api-access-2q4mr") pod "595377fa-b0c8-440a-a6f9-4bce1ce40352" (UID: "595377fa-b0c8-440a-a6f9-4bce1ce40352"). InnerVolumeSpecName "kube-api-access-2q4mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.013458 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-scripts" (OuterVolumeSpecName: "scripts") pod "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" (UID: "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.014718 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-kube-api-access-zhg5r" (OuterVolumeSpecName: "kube-api-access-zhg5r") pod "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" (UID: "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4"). InnerVolumeSpecName "kube-api-access-zhg5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.015009 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" (UID: "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.015758 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" (UID: "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.033915 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-config-data" (OuterVolumeSpecName: "config-data") pod "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" (UID: "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.033933 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" (UID: "c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.035982 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "595377fa-b0c8-440a-a6f9-4bce1ce40352" (UID: "595377fa-b0c8-440a-a6f9-4bce1ce40352"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.038709 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-config" (OuterVolumeSpecName: "config") pod "595377fa-b0c8-440a-a6f9-4bce1ce40352" (UID: "595377fa-b0c8-440a-a6f9-4bce1ce40352"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113487 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113542 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113564 4726 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113585 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhg5r\" (UniqueName: \"kubernetes.io/projected/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-kube-api-access-zhg5r\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113603 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113622 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595377fa-b0c8-440a-a6f9-4bce1ce40352-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113641 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113662 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q4mr\" (UniqueName: \"kubernetes.io/projected/595377fa-b0c8-440a-a6f9-4bce1ce40352-kube-api-access-2q4mr\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.113679 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.335636 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerStarted","Data":"fdd4b982192768220a2598051eb750b7ba0396e2789c1d26e90fcc0f6b78069c"} Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.354723 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gcxk9" event={"ID":"c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4","Type":"ContainerDied","Data":"fd08c571e7be6742f653f1d49da2b8eb5af3e46dbb011e29b95039489f364819"} Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.354765 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd08c571e7be6742f653f1d49da2b8eb5af3e46dbb011e29b95039489f364819" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.354813 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gcxk9" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.356803 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qk592" event={"ID":"0c3f7da5-afd6-44b1-b402-f3de8c39b518","Type":"ContainerDied","Data":"5d984993c31f6e5024c00c370fc01757bb75ed5bf1649bef2025fa584796daf3"} Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.356821 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d984993c31f6e5024c00c370fc01757bb75ed5bf1649bef2025fa584796daf3" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.356860 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qk592" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.366306 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4tm6n" event={"ID":"595377fa-b0c8-440a-a6f9-4bce1ce40352","Type":"ContainerDied","Data":"2bbcaa3e95542be9f390800249cb0388a83370bf11a1677f9dac3c529c440f61"} Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.366353 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bbcaa3e95542be9f390800249cb0388a83370bf11a1677f9dac3c529c440f61" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.366404 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4tm6n" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.420372 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b76988558-2489x"] Nov 23 20:25:23 crc kubenswrapper[4726]: E1123 20:25:23.420733 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" containerName="keystone-bootstrap" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.420750 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" containerName="keystone-bootstrap" Nov 23 20:25:23 crc kubenswrapper[4726]: E1123 20:25:23.420767 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="init" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.420773 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="init" Nov 23 20:25:23 crc kubenswrapper[4726]: E1123 20:25:23.420794 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c3f7da5-afd6-44b1-b402-f3de8c39b518" containerName="placement-db-sync" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.420801 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c3f7da5-afd6-44b1-b402-f3de8c39b518" containerName="placement-db-sync" Nov 23 20:25:23 crc kubenswrapper[4726]: E1123 20:25:23.420812 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="dnsmasq-dns" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.420818 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="dnsmasq-dns" Nov 23 20:25:23 crc kubenswrapper[4726]: E1123 20:25:23.420837 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595377fa-b0c8-440a-a6f9-4bce1ce40352" containerName="neutron-db-sync" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.420843 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="595377fa-b0c8-440a-a6f9-4bce1ce40352" containerName="neutron-db-sync" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.421002 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="595377fa-b0c8-440a-a6f9-4bce1ce40352" containerName="neutron-db-sync" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.421015 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="862ffc28-0d10-4190-9654-51ae9590043e" containerName="dnsmasq-dns" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.421023 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" containerName="keystone-bootstrap" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.421038 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c3f7da5-afd6-44b1-b402-f3de8c39b518" containerName="placement-db-sync" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.423497 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.426548 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.426679 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.426616 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.426855 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5dr84" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.427135 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.427165 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.445609 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b76988558-2489x"] Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.523452 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-credential-keys\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.523498 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-fernet-keys\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.523551 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw82p\" (UniqueName: \"kubernetes.io/projected/1071728a-1612-4dc4-9cbe-25c043998eb5-kube-api-access-jw82p\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.523714 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-combined-ca-bundle\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.523939 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-scripts\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.524045 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-config-data\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.524146 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-public-tls-certs\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.524225 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-internal-tls-certs\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.626152 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-scripts\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.626219 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-config-data\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.626242 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-public-tls-certs\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.626277 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-internal-tls-certs\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.626326 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-credential-keys\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.626345 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-fernet-keys\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.626368 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw82p\" (UniqueName: \"kubernetes.io/projected/1071728a-1612-4dc4-9cbe-25c043998eb5-kube-api-access-jw82p\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.626400 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-combined-ca-bundle\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.631936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-scripts\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.632327 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-credential-keys\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.633012 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-fernet-keys\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.633495 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-internal-tls-certs\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.642778 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-config-data\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.643265 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-combined-ca-bundle\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.651878 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1071728a-1612-4dc4-9cbe-25c043998eb5-public-tls-certs\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.651984 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw82p\" (UniqueName: \"kubernetes.io/projected/1071728a-1612-4dc4-9cbe-25c043998eb5-kube-api-access-jw82p\") pod \"keystone-b76988558-2489x\" (UID: \"1071728a-1612-4dc4-9cbe-25c043998eb5\") " pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.745061 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.845800 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-fbb56fbb6-7fzgk"] Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.847327 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.854228 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.854464 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.854634 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.854739 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f72c9" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.854841 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 23 20:25:23 crc kubenswrapper[4726]: I1123 20:25:23.912466 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fbb56fbb6-7fzgk"] Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.035808 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-public-tls-certs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.036343 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l2xp\" (UniqueName: \"kubernetes.io/projected/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-kube-api-access-7l2xp\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.036507 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-combined-ca-bundle\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.036575 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-scripts\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.036662 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-logs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.036726 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-config-data\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.036800 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-internal-tls-certs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.142775 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-combined-ca-bundle\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.142825 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-scripts\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.142887 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-logs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.142909 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-config-data\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.142936 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-internal-tls-certs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.142965 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-public-tls-certs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.142986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l2xp\" (UniqueName: \"kubernetes.io/projected/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-kube-api-access-7l2xp\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.146362 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-logs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.151812 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-config-data\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.158647 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-internal-tls-certs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.171280 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l2xp\" (UniqueName: \"kubernetes.io/projected/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-kube-api-access-7l2xp\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.177429 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-scripts\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.177838 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-public-tls-certs\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.179113 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d3aaf5-7d1c-46ac-8a80-5ab758b260d4-combined-ca-bundle\") pod \"placement-fbb56fbb6-7fzgk\" (UID: \"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4\") " pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.250191 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6677d66f85-t9772"] Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.253766 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.305675 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6677d66f85-t9772"] Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.349041 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-nb\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.349434 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-sb\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.349548 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-config\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.349681 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-dns-svc\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.349755 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4bpw\" (UniqueName: \"kubernetes.io/projected/1319d659-4cee-47fa-b083-370d5c905ded-kube-api-access-v4bpw\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.423728 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b76988558-2489x"] Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.450943 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-dns-svc\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.450980 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4bpw\" (UniqueName: \"kubernetes.io/projected/1319d659-4cee-47fa-b083-370d5c905ded-kube-api-access-v4bpw\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.451070 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-nb\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.451111 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-sb\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.451132 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-config\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.452271 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-dns-svc\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.452744 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-nb\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.452829 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-sb\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.453394 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-config\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.464611 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.476989 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c5d686b4-gshjj"] Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.478247 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4bpw\" (UniqueName: \"kubernetes.io/projected/1319d659-4cee-47fa-b083-370d5c905ded-kube-api-access-v4bpw\") pod \"dnsmasq-dns-6677d66f85-t9772\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.478717 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.483491 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.483655 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-hqg9v" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.484959 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.490180 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.502090 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c5d686b4-gshjj"] Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.553423 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-config\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.553511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-ovndb-tls-certs\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.553541 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-combined-ca-bundle\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.553591 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-httpd-config\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.553633 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rkls\" (UniqueName: \"kubernetes.io/projected/34abb904-d204-4941-afc7-21887ef6703c-kube-api-access-8rkls\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.598595 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.659783 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-ovndb-tls-certs\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.660374 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-combined-ca-bundle\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.660515 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-httpd-config\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.660622 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rkls\" (UniqueName: \"kubernetes.io/projected/34abb904-d204-4941-afc7-21887ef6703c-kube-api-access-8rkls\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.660707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-config\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.666468 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-config\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.685729 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-combined-ca-bundle\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.686444 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-ovndb-tls-certs\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.690089 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-httpd-config\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.696643 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rkls\" (UniqueName: \"kubernetes.io/projected/34abb904-d204-4941-afc7-21887ef6703c-kube-api-access-8rkls\") pod \"neutron-c5d686b4-gshjj\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:24 crc kubenswrapper[4726]: I1123 20:25:24.830084 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.028572 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fbb56fbb6-7fzgk"] Nov 23 20:25:25 crc kubenswrapper[4726]: W1123 20:25:25.038814 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71d3aaf5_7d1c_46ac_8a80_5ab758b260d4.slice/crio-6633266874e684abcdc7245cc025f0622226ea3c5afa635de1d5a9c4e203d6a3 WatchSource:0}: Error finding container 6633266874e684abcdc7245cc025f0622226ea3c5afa635de1d5a9c4e203d6a3: Status 404 returned error can't find the container with id 6633266874e684abcdc7245cc025f0622226ea3c5afa635de1d5a9c4e203d6a3 Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.263427 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6677d66f85-t9772"] Nov 23 20:25:25 crc kubenswrapper[4726]: W1123 20:25:25.278484 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1319d659_4cee_47fa_b083_370d5c905ded.slice/crio-05858fc5e47ed60ef6939cc456746a800533002ab0272924581c0b8fb455c1d8 WatchSource:0}: Error finding container 05858fc5e47ed60ef6939cc456746a800533002ab0272924581c0b8fb455c1d8: Status 404 returned error can't find the container with id 05858fc5e47ed60ef6939cc456746a800533002ab0272924581c0b8fb455c1d8 Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.327549 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c5d686b4-gshjj"] Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.416513 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6677d66f85-t9772" event={"ID":"1319d659-4cee-47fa-b083-370d5c905ded","Type":"ContainerStarted","Data":"05858fc5e47ed60ef6939cc456746a800533002ab0272924581c0b8fb455c1d8"} Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.420552 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b76988558-2489x" event={"ID":"1071728a-1612-4dc4-9cbe-25c043998eb5","Type":"ContainerStarted","Data":"c854b3ebf70fd1aecdfa806aacce89383dec9e6fa1a98b9061e16de4c17d49f5"} Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.420577 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b76988558-2489x" event={"ID":"1071728a-1612-4dc4-9cbe-25c043998eb5","Type":"ContainerStarted","Data":"fca47d4b2afe606db12c15fa2086a9838f3d509f3e2044e0ae6423020b0e81eb"} Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.421619 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.423957 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fbb56fbb6-7fzgk" event={"ID":"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4","Type":"ContainerStarted","Data":"1249fa97d68c52861de1c9f46f23cb990ee3d4ba39ba3b01bba1dda0f1724d1e"} Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.423985 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fbb56fbb6-7fzgk" event={"ID":"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4","Type":"ContainerStarted","Data":"6633266874e684abcdc7245cc025f0622226ea3c5afa635de1d5a9c4e203d6a3"} Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.428376 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c5d686b4-gshjj" event={"ID":"34abb904-d204-4941-afc7-21887ef6703c","Type":"ContainerStarted","Data":"c8ce54b856a9b1f7a597245b007458d38db72aada2788b847e730864b50c0f97"} Nov 23 20:25:25 crc kubenswrapper[4726]: I1123 20:25:25.439992 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-b76988558-2489x" podStartSLOduration=2.439953951 podStartE2EDuration="2.439953951s" podCreationTimestamp="2025-11-23 20:25:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:25.43847825 +0000 UTC m=+1033.587519206" watchObservedRunningTime="2025-11-23 20:25:25.439953951 +0000 UTC m=+1033.588994907" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.440624 4726 generic.go:334] "Generic (PLEG): container finished" podID="1319d659-4cee-47fa-b083-370d5c905ded" containerID="dafefbb4483a0071dfb868d640759eef23e4d2853e03b6f625e200a77c9418ef" exitCode=0 Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.440674 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6677d66f85-t9772" event={"ID":"1319d659-4cee-47fa-b083-370d5c905ded","Type":"ContainerDied","Data":"dafefbb4483a0071dfb868d640759eef23e4d2853e03b6f625e200a77c9418ef"} Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.446941 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s7rvr" event={"ID":"97c02ff5-89f0-4834-ab15-1d226f632e17","Type":"ContainerStarted","Data":"9e457b2e93890a3c0fe8b7a0d2d02dabbf8b6fe0371ca9884d190bb05668e590"} Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.452058 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fbb56fbb6-7fzgk" event={"ID":"71d3aaf5-7d1c-46ac-8a80-5ab758b260d4","Type":"ContainerStarted","Data":"38386be91fb99139a169a4824823ad2e31271252aa62540623ce2d437efb3756"} Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.452964 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.452997 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.461055 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c5d686b4-gshjj" event={"ID":"34abb904-d204-4941-afc7-21887ef6703c","Type":"ContainerStarted","Data":"5ba3211414beac95a6a7fbc5f238d09f393b54bf60c0601a3676dda0449a0a62"} Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.461137 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c5d686b4-gshjj" event={"ID":"34abb904-d204-4941-afc7-21887ef6703c","Type":"ContainerStarted","Data":"f375e027d360f6079f3519e5b5529c65a935a08a49b4be2f20daaa9e461a04d6"} Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.461184 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.531621 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-fbb56fbb6-7fzgk" podStartSLOduration=3.5315984 podStartE2EDuration="3.5315984s" podCreationTimestamp="2025-11-23 20:25:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:26.494526703 +0000 UTC m=+1034.643567669" watchObservedRunningTime="2025-11-23 20:25:26.5315984 +0000 UTC m=+1034.680639346" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.540635 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-s7rvr" podStartSLOduration=2.166390589 podStartE2EDuration="36.540615029s" podCreationTimestamp="2025-11-23 20:24:50 +0000 UTC" firstStartedPulling="2025-11-23 20:24:51.784095013 +0000 UTC m=+999.933135969" lastFinishedPulling="2025-11-23 20:25:26.158319453 +0000 UTC m=+1034.307360409" observedRunningTime="2025-11-23 20:25:26.5240357 +0000 UTC m=+1034.673076666" watchObservedRunningTime="2025-11-23 20:25:26.540615029 +0000 UTC m=+1034.689655985" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.896558 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c5d686b4-gshjj" podStartSLOduration=2.8965376259999998 podStartE2EDuration="2.896537626s" podCreationTimestamp="2025-11-23 20:25:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:26.565590751 +0000 UTC m=+1034.714631717" watchObservedRunningTime="2025-11-23 20:25:26.896537626 +0000 UTC m=+1035.045578582" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.899678 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-57cf6d944c-r5qgs"] Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.901112 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.910794 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.911014 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 23 20:25:26 crc kubenswrapper[4726]: I1123 20:25:26.911676 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57cf6d944c-r5qgs"] Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.027364 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-httpd-config\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.027439 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-combined-ca-bundle\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.027468 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-config\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.027691 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-public-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.027818 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv498\" (UniqueName: \"kubernetes.io/projected/6e680832-25f3-4328-bf10-35bd83ab52b9-kube-api-access-hv498\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.027909 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-internal-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.027999 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-ovndb-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.129951 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-internal-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.130016 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-ovndb-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.130052 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-httpd-config\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.130095 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-combined-ca-bundle\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.130124 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-config\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.130171 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-public-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.130206 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv498\" (UniqueName: \"kubernetes.io/projected/6e680832-25f3-4328-bf10-35bd83ab52b9-kube-api-access-hv498\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.135936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-public-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.137303 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-ovndb-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.137818 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-combined-ca-bundle\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.138305 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-httpd-config\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.139522 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-internal-tls-certs\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.155421 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6e680832-25f3-4328-bf10-35bd83ab52b9-config\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.167403 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv498\" (UniqueName: \"kubernetes.io/projected/6e680832-25f3-4328-bf10-35bd83ab52b9-kube-api-access-hv498\") pod \"neutron-57cf6d944c-r5qgs\" (UID: \"6e680832-25f3-4328-bf10-35bd83ab52b9\") " pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.220671 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.492393 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6677d66f85-t9772" event={"ID":"1319d659-4cee-47fa-b083-370d5c905ded","Type":"ContainerStarted","Data":"8c574ebef2f8231cab622d95f4c57f7abfbf1ed94f79a4917521f0bef03ddcf3"} Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.494238 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.523518 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6677d66f85-t9772" podStartSLOduration=3.523467866 podStartE2EDuration="3.523467866s" podCreationTimestamp="2025-11-23 20:25:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:27.515822555 +0000 UTC m=+1035.664863511" watchObservedRunningTime="2025-11-23 20:25:27.523467866 +0000 UTC m=+1035.672508812" Nov 23 20:25:27 crc kubenswrapper[4726]: I1123 20:25:27.824505 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57cf6d944c-r5qgs"] Nov 23 20:25:27 crc kubenswrapper[4726]: W1123 20:25:27.833487 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e680832_25f3_4328_bf10_35bd83ab52b9.slice/crio-0a24c811c9b74c15c1a9bbbdd632c714b9b6071351c6abf6880d474ab47c8276 WatchSource:0}: Error finding container 0a24c811c9b74c15c1a9bbbdd632c714b9b6071351c6abf6880d474ab47c8276: Status 404 returned error can't find the container with id 0a24c811c9b74c15c1a9bbbdd632c714b9b6071351c6abf6880d474ab47c8276 Nov 23 20:25:28 crc kubenswrapper[4726]: I1123 20:25:28.512121 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57cf6d944c-r5qgs" event={"ID":"6e680832-25f3-4328-bf10-35bd83ab52b9","Type":"ContainerStarted","Data":"39435684957a5a8d459f6824861d3132f274f9f4a97fa7401aac0042cb034313"} Nov 23 20:25:28 crc kubenswrapper[4726]: I1123 20:25:28.512725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57cf6d944c-r5qgs" event={"ID":"6e680832-25f3-4328-bf10-35bd83ab52b9","Type":"ContainerStarted","Data":"0a24c811c9b74c15c1a9bbbdd632c714b9b6071351c6abf6880d474ab47c8276"} Nov 23 20:25:29 crc kubenswrapper[4726]: I1123 20:25:29.524597 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fq2c6" event={"ID":"b3d6b028-7e7a-4c1d-a05e-5f49976b823d","Type":"ContainerStarted","Data":"603e083de46ca863689b365fdd8a22dfeba96b03369b30aa6e6bb0ca213e1e79"} Nov 23 20:25:29 crc kubenswrapper[4726]: I1123 20:25:29.527679 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57cf6d944c-r5qgs" event={"ID":"6e680832-25f3-4328-bf10-35bd83ab52b9","Type":"ContainerStarted","Data":"28f8e580af640195483603d4684861ea16bcee2bf2bc68a1344c9d43064928af"} Nov 23 20:25:29 crc kubenswrapper[4726]: I1123 20:25:29.527774 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:29 crc kubenswrapper[4726]: I1123 20:25:29.530454 4726 generic.go:334] "Generic (PLEG): container finished" podID="97c02ff5-89f0-4834-ab15-1d226f632e17" containerID="9e457b2e93890a3c0fe8b7a0d2d02dabbf8b6fe0371ca9884d190bb05668e590" exitCode=0 Nov 23 20:25:29 crc kubenswrapper[4726]: I1123 20:25:29.530483 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s7rvr" event={"ID":"97c02ff5-89f0-4834-ab15-1d226f632e17","Type":"ContainerDied","Data":"9e457b2e93890a3c0fe8b7a0d2d02dabbf8b6fe0371ca9884d190bb05668e590"} Nov 23 20:25:29 crc kubenswrapper[4726]: I1123 20:25:29.549906 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-fq2c6" podStartSLOduration=3.358554661 podStartE2EDuration="39.549853209s" podCreationTimestamp="2025-11-23 20:24:50 +0000 UTC" firstStartedPulling="2025-11-23 20:24:52.034917569 +0000 UTC m=+1000.183958525" lastFinishedPulling="2025-11-23 20:25:28.226216117 +0000 UTC m=+1036.375257073" observedRunningTime="2025-11-23 20:25:29.540655835 +0000 UTC m=+1037.689696811" watchObservedRunningTime="2025-11-23 20:25:29.549853209 +0000 UTC m=+1037.698894185" Nov 23 20:25:29 crc kubenswrapper[4726]: I1123 20:25:29.587635 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-57cf6d944c-r5qgs" podStartSLOduration=3.587612615 podStartE2EDuration="3.587612615s" podCreationTimestamp="2025-11-23 20:25:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:29.583030138 +0000 UTC m=+1037.732071134" watchObservedRunningTime="2025-11-23 20:25:29.587612615 +0000 UTC m=+1037.736653581" Nov 23 20:25:32 crc kubenswrapper[4726]: I1123 20:25:32.815440 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:25:32 crc kubenswrapper[4726]: I1123 20:25:32.982735 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-combined-ca-bundle\") pod \"97c02ff5-89f0-4834-ab15-1d226f632e17\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " Nov 23 20:25:32 crc kubenswrapper[4726]: I1123 20:25:32.982830 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-db-sync-config-data\") pod \"97c02ff5-89f0-4834-ab15-1d226f632e17\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " Nov 23 20:25:32 crc kubenswrapper[4726]: I1123 20:25:32.982977 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tm4q\" (UniqueName: \"kubernetes.io/projected/97c02ff5-89f0-4834-ab15-1d226f632e17-kube-api-access-5tm4q\") pod \"97c02ff5-89f0-4834-ab15-1d226f632e17\" (UID: \"97c02ff5-89f0-4834-ab15-1d226f632e17\") " Nov 23 20:25:32 crc kubenswrapper[4726]: I1123 20:25:32.989902 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97c02ff5-89f0-4834-ab15-1d226f632e17-kube-api-access-5tm4q" (OuterVolumeSpecName: "kube-api-access-5tm4q") pod "97c02ff5-89f0-4834-ab15-1d226f632e17" (UID: "97c02ff5-89f0-4834-ab15-1d226f632e17"). InnerVolumeSpecName "kube-api-access-5tm4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:32 crc kubenswrapper[4726]: I1123 20:25:32.996704 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "97c02ff5-89f0-4834-ab15-1d226f632e17" (UID: "97c02ff5-89f0-4834-ab15-1d226f632e17"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.015128 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97c02ff5-89f0-4834-ab15-1d226f632e17" (UID: "97c02ff5-89f0-4834-ab15-1d226f632e17"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.084510 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.084540 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tm4q\" (UniqueName: \"kubernetes.io/projected/97c02ff5-89f0-4834-ab15-1d226f632e17-kube-api-access-5tm4q\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.084550 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c02ff5-89f0-4834-ab15-1d226f632e17-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.566255 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerStarted","Data":"77c17bb00fe8203cb0bded7db33549111b5f33c679b58deefd29ae1d8baa3c32"} Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.566936 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.566365 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="ceilometer-central-agent" containerID="cri-o://27d51503001e6dc16b2018d4a060cc21f99e42ec9d6537683ff985b7654e9697" gracePeriod=30 Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.566451 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="ceilometer-notification-agent" containerID="cri-o://48f4363e2fa89c5a8408a6a7c21589d4c82008c2e40bf3f98d946ad9dc5b6304" gracePeriod=30 Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.566512 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="proxy-httpd" containerID="cri-o://77c17bb00fe8203cb0bded7db33549111b5f33c679b58deefd29ae1d8baa3c32" gracePeriod=30 Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.566412 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="sg-core" containerID="cri-o://fdd4b982192768220a2598051eb750b7ba0396e2789c1d26e90fcc0f6b78069c" gracePeriod=30 Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.569272 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s7rvr" event={"ID":"97c02ff5-89f0-4834-ab15-1d226f632e17","Type":"ContainerDied","Data":"0c830b3579757cc5270044eba0092933d63b36e0784c2906e1fecfec26cd7fab"} Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.569431 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c830b3579757cc5270044eba0092933d63b36e0784c2906e1fecfec26cd7fab" Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.569587 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s7rvr" Nov 23 20:25:33 crc kubenswrapper[4726]: I1123 20:25:33.852146 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.007409222 podStartE2EDuration="43.852128542s" podCreationTimestamp="2025-11-23 20:24:50 +0000 UTC" firstStartedPulling="2025-11-23 20:24:51.561909271 +0000 UTC m=+999.710950217" lastFinishedPulling="2025-11-23 20:25:33.406628581 +0000 UTC m=+1041.555669537" observedRunningTime="2025-11-23 20:25:33.596987983 +0000 UTC m=+1041.746028949" watchObservedRunningTime="2025-11-23 20:25:33.852128542 +0000 UTC m=+1042.001169488" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.153115 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7b6bd669d5-fsxlx"] Nov 23 20:25:34 crc kubenswrapper[4726]: E1123 20:25:34.153895 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c02ff5-89f0-4834-ab15-1d226f632e17" containerName="barbican-db-sync" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.153915 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c02ff5-89f0-4834-ab15-1d226f632e17" containerName="barbican-db-sync" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.154217 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="97c02ff5-89f0-4834-ab15-1d226f632e17" containerName="barbican-db-sync" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.155393 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.168361 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-v2kht" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.168407 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.168974 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.194507 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b6bd669d5-fsxlx"] Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.237374 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-644b66b48-csfsq"] Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.239411 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.242141 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.283448 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-644b66b48-csfsq"] Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.304477 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3041645e-f5e5-4faf-b2de-365bcd156177-logs\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.304725 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6rfw\" (UniqueName: \"kubernetes.io/projected/3041645e-f5e5-4faf-b2de-365bcd156177-kube-api-access-s6rfw\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.304823 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-combined-ca-bundle\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.304935 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-combined-ca-bundle\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.305023 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-config-data\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.305099 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-config-data-custom\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.305199 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbkwm\" (UniqueName: \"kubernetes.io/projected/02190dad-4d68-4cc1-937c-a45b260b4a95-kube-api-access-nbkwm\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.305276 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-config-data\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.305349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02190dad-4d68-4cc1-937c-a45b260b4a95-logs\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.305425 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-config-data-custom\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.324504 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6677d66f85-t9772"] Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.324750 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6677d66f85-t9772" podUID="1319d659-4cee-47fa-b083-370d5c905ded" containerName="dnsmasq-dns" containerID="cri-o://8c574ebef2f8231cab622d95f4c57f7abfbf1ed94f79a4917521f0bef03ddcf3" gracePeriod=10 Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.328306 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.340267 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-r52hj"] Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.341755 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.362083 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-r52hj"] Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407360 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6rfw\" (UniqueName: \"kubernetes.io/projected/3041645e-f5e5-4faf-b2de-365bcd156177-kube-api-access-s6rfw\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407431 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-combined-ca-bundle\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407480 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-combined-ca-bundle\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407505 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-config-data\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407525 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-config-data-custom\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407553 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbkwm\" (UniqueName: \"kubernetes.io/projected/02190dad-4d68-4cc1-937c-a45b260b4a95-kube-api-access-nbkwm\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407579 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-config-data\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407598 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02190dad-4d68-4cc1-937c-a45b260b4a95-logs\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407618 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-config-data-custom\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.407649 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3041645e-f5e5-4faf-b2de-365bcd156177-logs\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.408121 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3041645e-f5e5-4faf-b2de-365bcd156177-logs\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.412513 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02190dad-4d68-4cc1-937c-a45b260b4a95-logs\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.415844 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-config-data-custom\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.417836 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-config-data\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.436539 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-combined-ca-bundle\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.437783 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02190dad-4d68-4cc1-937c-a45b260b4a95-config-data\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.438642 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-combined-ca-bundle\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.465475 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-595b84cd88-8zlj4"] Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.474190 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.485461 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3041645e-f5e5-4faf-b2de-365bcd156177-config-data-custom\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.485513 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.488200 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-595b84cd88-8zlj4"] Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.503479 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6rfw\" (UniqueName: \"kubernetes.io/projected/3041645e-f5e5-4faf-b2de-365bcd156177-kube-api-access-s6rfw\") pod \"barbican-worker-7b6bd669d5-fsxlx\" (UID: \"3041645e-f5e5-4faf-b2de-365bcd156177\") " pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.503889 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbkwm\" (UniqueName: \"kubernetes.io/projected/02190dad-4d68-4cc1-937c-a45b260b4a95-kube-api-access-nbkwm\") pod \"barbican-keystone-listener-644b66b48-csfsq\" (UID: \"02190dad-4d68-4cc1-937c-a45b260b4a95\") " pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.518466 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.518816 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtv5p\" (UniqueName: \"kubernetes.io/projected/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-kube-api-access-jtv5p\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.518955 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-config\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.519065 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-sb\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.519183 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data-custom\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.519283 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-nb\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.519376 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-dns-svc\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.519451 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgcvk\" (UniqueName: \"kubernetes.io/projected/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-kube-api-access-jgcvk\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.519525 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-combined-ca-bundle\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.519600 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-logs\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.565438 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-644b66b48-csfsq" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.593344 4726 generic.go:334] "Generic (PLEG): container finished" podID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerID="fdd4b982192768220a2598051eb750b7ba0396e2789c1d26e90fcc0f6b78069c" exitCode=2 Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.593371 4726 generic.go:334] "Generic (PLEG): container finished" podID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerID="27d51503001e6dc16b2018d4a060cc21f99e42ec9d6537683ff985b7654e9697" exitCode=0 Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.600253 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6677d66f85-t9772" podUID="1319d659-4cee-47fa-b083-370d5c905ded" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: connect: connection refused" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.603720 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerDied","Data":"fdd4b982192768220a2598051eb750b7ba0396e2789c1d26e90fcc0f6b78069c"} Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.603744 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerDied","Data":"27d51503001e6dc16b2018d4a060cc21f99e42ec9d6537683ff985b7654e9697"} Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.603830 4726 generic.go:334] "Generic (PLEG): container finished" podID="1319d659-4cee-47fa-b083-370d5c905ded" containerID="8c574ebef2f8231cab622d95f4c57f7abfbf1ed94f79a4917521f0bef03ddcf3" exitCode=0 Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.603848 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6677d66f85-t9772" event={"ID":"1319d659-4cee-47fa-b083-370d5c905ded","Type":"ContainerDied","Data":"8c574ebef2f8231cab622d95f4c57f7abfbf1ed94f79a4917521f0bef03ddcf3"} Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622353 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtv5p\" (UniqueName: \"kubernetes.io/projected/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-kube-api-access-jtv5p\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622407 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-config\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622428 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-sb\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622458 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data-custom\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622476 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-nb\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622516 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-dns-svc\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622543 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgcvk\" (UniqueName: \"kubernetes.io/projected/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-kube-api-access-jgcvk\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622565 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-combined-ca-bundle\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622592 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-logs\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.622622 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.623500 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-config\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.624216 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-sb\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.625085 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-nb\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.625373 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-dns-svc\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.625750 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-logs\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.632617 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-combined-ca-bundle\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.634181 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.635849 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data-custom\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.654661 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgcvk\" (UniqueName: \"kubernetes.io/projected/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-kube-api-access-jgcvk\") pod \"dnsmasq-dns-844b557b9c-r52hj\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.661422 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.670826 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtv5p\" (UniqueName: \"kubernetes.io/projected/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-kube-api-access-jtv5p\") pod \"barbican-api-595b84cd88-8zlj4\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.782437 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b6bd669d5-fsxlx" Nov 23 20:25:34 crc kubenswrapper[4726]: I1123 20:25:34.884463 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.005672 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.051365 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-nb\") pod \"1319d659-4cee-47fa-b083-370d5c905ded\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.051452 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-config\") pod \"1319d659-4cee-47fa-b083-370d5c905ded\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.051538 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-dns-svc\") pod \"1319d659-4cee-47fa-b083-370d5c905ded\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.051572 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4bpw\" (UniqueName: \"kubernetes.io/projected/1319d659-4cee-47fa-b083-370d5c905ded-kube-api-access-v4bpw\") pod \"1319d659-4cee-47fa-b083-370d5c905ded\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.051666 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-sb\") pod \"1319d659-4cee-47fa-b083-370d5c905ded\" (UID: \"1319d659-4cee-47fa-b083-370d5c905ded\") " Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.066784 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1319d659-4cee-47fa-b083-370d5c905ded-kube-api-access-v4bpw" (OuterVolumeSpecName: "kube-api-access-v4bpw") pod "1319d659-4cee-47fa-b083-370d5c905ded" (UID: "1319d659-4cee-47fa-b083-370d5c905ded"). InnerVolumeSpecName "kube-api-access-v4bpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.154785 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4bpw\" (UniqueName: \"kubernetes.io/projected/1319d659-4cee-47fa-b083-370d5c905ded-kube-api-access-v4bpw\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.165464 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1319d659-4cee-47fa-b083-370d5c905ded" (UID: "1319d659-4cee-47fa-b083-370d5c905ded"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.177668 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1319d659-4cee-47fa-b083-370d5c905ded" (UID: "1319d659-4cee-47fa-b083-370d5c905ded"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.206487 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-config" (OuterVolumeSpecName: "config") pod "1319d659-4cee-47fa-b083-370d5c905ded" (UID: "1319d659-4cee-47fa-b083-370d5c905ded"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.227354 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1319d659-4cee-47fa-b083-370d5c905ded" (UID: "1319d659-4cee-47fa-b083-370d5c905ded"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.266790 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.266823 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.266833 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.266843 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1319d659-4cee-47fa-b083-370d5c905ded-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.293671 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-644b66b48-csfsq"] Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.437428 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-r52hj"] Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.662736 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b6bd669d5-fsxlx"] Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.668599 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-644b66b48-csfsq" event={"ID":"02190dad-4d68-4cc1-937c-a45b260b4a95","Type":"ContainerStarted","Data":"4a27ac512c578e122a9ba648d4ca15cd975b5e70601337581bd1d7a108d1677f"} Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.707545 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6677d66f85-t9772" event={"ID":"1319d659-4cee-47fa-b083-370d5c905ded","Type":"ContainerDied","Data":"05858fc5e47ed60ef6939cc456746a800533002ab0272924581c0b8fb455c1d8"} Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.707594 4726 scope.go:117] "RemoveContainer" containerID="8c574ebef2f8231cab622d95f4c57f7abfbf1ed94f79a4917521f0bef03ddcf3" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.707719 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6677d66f85-t9772" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.712602 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" event={"ID":"e6db9abd-d34a-4628-9073-f76c5ee4ea7a","Type":"ContainerStarted","Data":"e0590802f295cf0f163ebc6032b626cb7c28ca28326c50ce24431a44cb1fb853"} Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.750613 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-595b84cd88-8zlj4"] Nov 23 20:25:35 crc kubenswrapper[4726]: W1123 20:25:35.800247 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a95fa41_6d0c_44e2_8e87_2045ef8bc345.slice/crio-7c32aa99bc85a3f27774f3d6a189973f74993a62cf08bbc91716ccc5e3e62606 WatchSource:0}: Error finding container 7c32aa99bc85a3f27774f3d6a189973f74993a62cf08bbc91716ccc5e3e62606: Status 404 returned error can't find the container with id 7c32aa99bc85a3f27774f3d6a189973f74993a62cf08bbc91716ccc5e3e62606 Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.807550 4726 scope.go:117] "RemoveContainer" containerID="dafefbb4483a0071dfb868d640759eef23e4d2853e03b6f625e200a77c9418ef" Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.828852 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6677d66f85-t9772"] Nov 23 20:25:35 crc kubenswrapper[4726]: I1123 20:25:35.844033 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6677d66f85-t9772"] Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.619707 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1319d659-4cee-47fa-b083-370d5c905ded" path="/var/lib/kubelet/pods/1319d659-4cee-47fa-b083-370d5c905ded/volumes" Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.725038 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-595b84cd88-8zlj4" event={"ID":"9a95fa41-6d0c-44e2-8e87-2045ef8bc345","Type":"ContainerStarted","Data":"e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3"} Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.725085 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-595b84cd88-8zlj4" event={"ID":"9a95fa41-6d0c-44e2-8e87-2045ef8bc345","Type":"ContainerStarted","Data":"4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33"} Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.725098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-595b84cd88-8zlj4" event={"ID":"9a95fa41-6d0c-44e2-8e87-2045ef8bc345","Type":"ContainerStarted","Data":"7c32aa99bc85a3f27774f3d6a189973f74993a62cf08bbc91716ccc5e3e62606"} Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.725146 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.725168 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.728590 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b6bd669d5-fsxlx" event={"ID":"3041645e-f5e5-4faf-b2de-365bcd156177","Type":"ContainerStarted","Data":"9485654ed93488671286aedd695deedd150145c663bfdde2169a3f05c941f7c6"} Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.731147 4726 generic.go:334] "Generic (PLEG): container finished" podID="b3d6b028-7e7a-4c1d-a05e-5f49976b823d" containerID="603e083de46ca863689b365fdd8a22dfeba96b03369b30aa6e6bb0ca213e1e79" exitCode=0 Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.731230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fq2c6" event={"ID":"b3d6b028-7e7a-4c1d-a05e-5f49976b823d","Type":"ContainerDied","Data":"603e083de46ca863689b365fdd8a22dfeba96b03369b30aa6e6bb0ca213e1e79"} Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.735607 4726 generic.go:334] "Generic (PLEG): container finished" podID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" containerID="c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53" exitCode=0 Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.735634 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" event={"ID":"e6db9abd-d34a-4628-9073-f76c5ee4ea7a","Type":"ContainerDied","Data":"c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53"} Nov 23 20:25:36 crc kubenswrapper[4726]: I1123 20:25:36.753275 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-595b84cd88-8zlj4" podStartSLOduration=2.753259311 podStartE2EDuration="2.753259311s" podCreationTimestamp="2025-11-23 20:25:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:36.74833227 +0000 UTC m=+1044.897373226" watchObservedRunningTime="2025-11-23 20:25:36.753259311 +0000 UTC m=+1044.902300267" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.675993 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-85855866bd-77958"] Nov 23 20:25:37 crc kubenswrapper[4726]: E1123 20:25:37.676656 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1319d659-4cee-47fa-b083-370d5c905ded" containerName="dnsmasq-dns" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.676673 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1319d659-4cee-47fa-b083-370d5c905ded" containerName="dnsmasq-dns" Nov 23 20:25:37 crc kubenswrapper[4726]: E1123 20:25:37.676691 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1319d659-4cee-47fa-b083-370d5c905ded" containerName="init" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.676699 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1319d659-4cee-47fa-b083-370d5c905ded" containerName="init" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.676913 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1319d659-4cee-47fa-b083-370d5c905ded" containerName="dnsmasq-dns" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.677837 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.680341 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.680786 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.699808 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85855866bd-77958"] Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.781888 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-config-data\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.782145 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-internal-tls-certs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.782331 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-combined-ca-bundle\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.782416 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-public-tls-certs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.782505 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-config-data-custom\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.782587 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cddea6c0-9834-4a47-b30c-ef32194ef6ea-logs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.782673 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x48k9\" (UniqueName: \"kubernetes.io/projected/cddea6c0-9834-4a47-b30c-ef32194ef6ea-kube-api-access-x48k9\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.884808 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x48k9\" (UniqueName: \"kubernetes.io/projected/cddea6c0-9834-4a47-b30c-ef32194ef6ea-kube-api-access-x48k9\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.884863 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-config-data\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.884943 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-internal-tls-certs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.884993 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-combined-ca-bundle\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.885025 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-public-tls-certs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.885054 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-config-data-custom\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.885091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cddea6c0-9834-4a47-b30c-ef32194ef6ea-logs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.885396 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cddea6c0-9834-4a47-b30c-ef32194ef6ea-logs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.890808 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-internal-tls-certs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.893017 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-combined-ca-bundle\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.901258 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x48k9\" (UniqueName: \"kubernetes.io/projected/cddea6c0-9834-4a47-b30c-ef32194ef6ea-kube-api-access-x48k9\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.909660 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-public-tls-certs\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.910127 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-config-data-custom\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.916221 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cddea6c0-9834-4a47-b30c-ef32194ef6ea-config-data\") pod \"barbican-api-85855866bd-77958\" (UID: \"cddea6c0-9834-4a47-b30c-ef32194ef6ea\") " pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:37 crc kubenswrapper[4726]: I1123 20:25:37.995204 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.196734 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.298446 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-config-data\") pod \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.298633 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-combined-ca-bundle\") pod \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.298699 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x95f6\" (UniqueName: \"kubernetes.io/projected/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-kube-api-access-x95f6\") pod \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.298725 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-db-sync-config-data\") pod \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.298769 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-scripts\") pod \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.298794 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-etc-machine-id\") pod \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\" (UID: \"b3d6b028-7e7a-4c1d-a05e-5f49976b823d\") " Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.299239 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b3d6b028-7e7a-4c1d-a05e-5f49976b823d" (UID: "b3d6b028-7e7a-4c1d-a05e-5f49976b823d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.309729 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-scripts" (OuterVolumeSpecName: "scripts") pod "b3d6b028-7e7a-4c1d-a05e-5f49976b823d" (UID: "b3d6b028-7e7a-4c1d-a05e-5f49976b823d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.318134 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b3d6b028-7e7a-4c1d-a05e-5f49976b823d" (UID: "b3d6b028-7e7a-4c1d-a05e-5f49976b823d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.333489 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-kube-api-access-x95f6" (OuterVolumeSpecName: "kube-api-access-x95f6") pod "b3d6b028-7e7a-4c1d-a05e-5f49976b823d" (UID: "b3d6b028-7e7a-4c1d-a05e-5f49976b823d"). InnerVolumeSpecName "kube-api-access-x95f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.390279 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3d6b028-7e7a-4c1d-a05e-5f49976b823d" (UID: "b3d6b028-7e7a-4c1d-a05e-5f49976b823d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.400621 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.400788 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x95f6\" (UniqueName: \"kubernetes.io/projected/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-kube-api-access-x95f6\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.400960 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.401019 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.401075 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.423127 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-config-data" (OuterVolumeSpecName: "config-data") pod "b3d6b028-7e7a-4c1d-a05e-5f49976b823d" (UID: "b3d6b028-7e7a-4c1d-a05e-5f49976b823d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.502215 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3d6b028-7e7a-4c1d-a05e-5f49976b823d-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.512114 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85855866bd-77958"] Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.759752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" event={"ID":"e6db9abd-d34a-4628-9073-f76c5ee4ea7a","Type":"ContainerStarted","Data":"a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.759933 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.761478 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85855866bd-77958" event={"ID":"cddea6c0-9834-4a47-b30c-ef32194ef6ea","Type":"ContainerStarted","Data":"b5e73a9a5734da64f4c153a36ffee67996dd4e275202d8d307a1179037a93bba"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.761516 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85855866bd-77958" event={"ID":"cddea6c0-9834-4a47-b30c-ef32194ef6ea","Type":"ContainerStarted","Data":"4bd4b50d0e2b31966ad7684ac81521d179dbcf1f22a0b8c4e7e63009267e44fb"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.770157 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b6bd669d5-fsxlx" event={"ID":"3041645e-f5e5-4faf-b2de-365bcd156177","Type":"ContainerStarted","Data":"fb1a68125a0542bbab3f1936620d44f2cf799271337ddf969f634306baf1f0b7"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.770232 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b6bd669d5-fsxlx" event={"ID":"3041645e-f5e5-4faf-b2de-365bcd156177","Type":"ContainerStarted","Data":"b1cd8b43b60dcb17b41478b3206154ae5c19fa4c19f6267eae84c9b499869d36"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.784296 4726 generic.go:334] "Generic (PLEG): container finished" podID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerID="48f4363e2fa89c5a8408a6a7c21589d4c82008c2e40bf3f98d946ad9dc5b6304" exitCode=0 Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.784370 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerDied","Data":"48f4363e2fa89c5a8408a6a7c21589d4c82008c2e40bf3f98d946ad9dc5b6304"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.786670 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" podStartSLOduration=4.786654507 podStartE2EDuration="4.786654507s" podCreationTimestamp="2025-11-23 20:25:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:38.7762571 +0000 UTC m=+1046.925298056" watchObservedRunningTime="2025-11-23 20:25:38.786654507 +0000 UTC m=+1046.935695463" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.789182 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fq2c6" event={"ID":"b3d6b028-7e7a-4c1d-a05e-5f49976b823d","Type":"ContainerDied","Data":"678c2b73fdba7928193bbeee4dc96ac997490c6b90bc23b4696ea88f13ebf102"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.789214 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="678c2b73fdba7928193bbeee4dc96ac997490c6b90bc23b4696ea88f13ebf102" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.789293 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fq2c6" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.793717 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-644b66b48-csfsq" event={"ID":"02190dad-4d68-4cc1-937c-a45b260b4a95","Type":"ContainerStarted","Data":"de3c5aee0b66ffcd382ed3c987625bf985d23d5a86abbbb944d2a1aa935887f3"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.794080 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-644b66b48-csfsq" event={"ID":"02190dad-4d68-4cc1-937c-a45b260b4a95","Type":"ContainerStarted","Data":"176e129de91524bba0c66b5d25162ad275f52c630141204eff2edd5a1fdfb4db"} Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.800278 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7b6bd669d5-fsxlx" podStartSLOduration=3.003899211 podStartE2EDuration="4.80026605s" podCreationTimestamp="2025-11-23 20:25:34 +0000 UTC" firstStartedPulling="2025-11-23 20:25:35.696574253 +0000 UTC m=+1043.845615209" lastFinishedPulling="2025-11-23 20:25:37.492941092 +0000 UTC m=+1045.641982048" observedRunningTime="2025-11-23 20:25:38.798236246 +0000 UTC m=+1046.947277212" watchObservedRunningTime="2025-11-23 20:25:38.80026605 +0000 UTC m=+1046.949307006" Nov 23 20:25:38 crc kubenswrapper[4726]: I1123 20:25:38.821533 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-644b66b48-csfsq" podStartSLOduration=2.603709647 podStartE2EDuration="4.821512976s" podCreationTimestamp="2025-11-23 20:25:34 +0000 UTC" firstStartedPulling="2025-11-23 20:25:35.27501002 +0000 UTC m=+1043.424050976" lastFinishedPulling="2025-11-23 20:25:37.492813349 +0000 UTC m=+1045.641854305" observedRunningTime="2025-11-23 20:25:38.819903953 +0000 UTC m=+1046.968944939" watchObservedRunningTime="2025-11-23 20:25:38.821512976 +0000 UTC m=+1046.970553932" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.089332 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:39 crc kubenswrapper[4726]: E1123 20:25:39.089821 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d6b028-7e7a-4c1d-a05e-5f49976b823d" containerName="cinder-db-sync" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.089844 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d6b028-7e7a-4c1d-a05e-5f49976b823d" containerName="cinder-db-sync" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.090083 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d6b028-7e7a-4c1d-a05e-5f49976b823d" containerName="cinder-db-sync" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.091072 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.093522 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.094712 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-fgrl4" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.095082 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.098193 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.183696 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.221963 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.222018 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.222064 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pskwm\" (UniqueName: \"kubernetes.io/projected/67dea71d-9dd9-4945-812e-df9a679aa505-kube-api-access-pskwm\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.222138 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.222164 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67dea71d-9dd9-4945-812e-df9a679aa505-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.222210 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-scripts\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.243159 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-r52hj"] Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.280225 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-775457b975-ppzwh"] Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.282273 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.298620 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-775457b975-ppzwh"] Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.329883 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-scripts\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.329984 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-config\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330018 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-dns-svc\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330068 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-sb\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330096 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330170 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn227\" (UniqueName: \"kubernetes.io/projected/b5b50700-1688-4d3f-8508-30a069c33ac7-kube-api-access-vn227\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pskwm\" (UniqueName: \"kubernetes.io/projected/67dea71d-9dd9-4945-812e-df9a679aa505-kube-api-access-pskwm\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330226 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330244 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-nb\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330271 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67dea71d-9dd9-4945-812e-df9a679aa505-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.330360 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67dea71d-9dd9-4945-812e-df9a679aa505-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.339132 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-scripts\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.342350 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.371587 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.378811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.383503 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pskwm\" (UniqueName: \"kubernetes.io/projected/67dea71d-9dd9-4945-812e-df9a679aa505-kube-api-access-pskwm\") pod \"cinder-scheduler-0\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.409551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.411119 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.426732 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.429816 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.436422 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-config\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.436465 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-dns-svc\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.436512 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-sb\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.436553 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn227\" (UniqueName: \"kubernetes.io/projected/b5b50700-1688-4d3f-8508-30a069c33ac7-kube-api-access-vn227\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.436588 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-nb\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.437392 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-nb\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.437904 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-config\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.438408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-dns-svc\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.458204 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-sb\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.486018 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.504620 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn227\" (UniqueName: \"kubernetes.io/projected/b5b50700-1688-4d3f-8508-30a069c33ac7-kube-api-access-vn227\") pod \"dnsmasq-dns-775457b975-ppzwh\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.543433 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb831c8-3f64-44d9-b867-be31f64d4388-logs\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.543501 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnlxv\" (UniqueName: \"kubernetes.io/projected/8cb831c8-3f64-44d9-b867-be31f64d4388-kube-api-access-pnlxv\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.543580 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb831c8-3f64-44d9-b867-be31f64d4388-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.543614 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.543633 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-scripts\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.543670 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data-custom\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.543687 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.605845 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.646935 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb831c8-3f64-44d9-b867-be31f64d4388-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.646864 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb831c8-3f64-44d9-b867-be31f64d4388-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.647026 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.647044 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-scripts\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.647075 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data-custom\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.647095 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.647143 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb831c8-3f64-44d9-b867-be31f64d4388-logs\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.647180 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnlxv\" (UniqueName: \"kubernetes.io/projected/8cb831c8-3f64-44d9-b867-be31f64d4388-kube-api-access-pnlxv\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.648586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb831c8-3f64-44d9-b867-be31f64d4388-logs\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.651048 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data-custom\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.659646 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-scripts\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.660452 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.662032 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.673434 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnlxv\" (UniqueName: \"kubernetes.io/projected/8cb831c8-3f64-44d9-b867-be31f64d4388-kube-api-access-pnlxv\") pod \"cinder-api-0\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.785284 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.825053 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85855866bd-77958" event={"ID":"cddea6c0-9834-4a47-b30c-ef32194ef6ea","Type":"ContainerStarted","Data":"6c1cbb872694523d1d54bedf6d4e587843fa2eba1d16c8e9b2424d785fae0340"} Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.825525 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.825581 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:39 crc kubenswrapper[4726]: I1123 20:25:39.860091 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-85855866bd-77958" podStartSLOduration=2.860074601 podStartE2EDuration="2.860074601s" podCreationTimestamp="2025-11-23 20:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:39.859952878 +0000 UTC m=+1048.008993844" watchObservedRunningTime="2025-11-23 20:25:39.860074601 +0000 UTC m=+1048.009115557" Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.037835 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.258668 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-775457b975-ppzwh"] Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.566127 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.840565 4726 generic.go:334] "Generic (PLEG): container finished" podID="b5b50700-1688-4d3f-8508-30a069c33ac7" containerID="c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5" exitCode=0 Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.840660 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-ppzwh" event={"ID":"b5b50700-1688-4d3f-8508-30a069c33ac7","Type":"ContainerDied","Data":"c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5"} Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.840707 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-ppzwh" event={"ID":"b5b50700-1688-4d3f-8508-30a069c33ac7","Type":"ContainerStarted","Data":"0521ca618178a57a352b7b7474947aa0b0fca73f3e03df21cf60df5ba443b89e"} Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.846082 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb831c8-3f64-44d9-b867-be31f64d4388","Type":"ContainerStarted","Data":"c74f09073460d9b3b88acd7ff45817a3e462027cb2b4b2d9547279c4c9a892af"} Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.855272 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" podUID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" containerName="dnsmasq-dns" containerID="cri-o://a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5" gracePeriod=10 Nov 23 20:25:40 crc kubenswrapper[4726]: I1123 20:25:40.855646 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67dea71d-9dd9-4945-812e-df9a679aa505","Type":"ContainerStarted","Data":"49179fca25bde95843efa7ecee03089e80e13529f9b5a842c3f7295a861be3a2"} Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.475348 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.600892 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-config\") pod \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.601221 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-sb\") pod \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.601272 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-dns-svc\") pod \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.601314 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-nb\") pod \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.601390 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgcvk\" (UniqueName: \"kubernetes.io/projected/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-kube-api-access-jgcvk\") pod \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\" (UID: \"e6db9abd-d34a-4628-9073-f76c5ee4ea7a\") " Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.616110 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-kube-api-access-jgcvk" (OuterVolumeSpecName: "kube-api-access-jgcvk") pod "e6db9abd-d34a-4628-9073-f76c5ee4ea7a" (UID: "e6db9abd-d34a-4628-9073-f76c5ee4ea7a"). InnerVolumeSpecName "kube-api-access-jgcvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.703379 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgcvk\" (UniqueName: \"kubernetes.io/projected/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-kube-api-access-jgcvk\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.734143 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e6db9abd-d34a-4628-9073-f76c5ee4ea7a" (UID: "e6db9abd-d34a-4628-9073-f76c5ee4ea7a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.734830 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e6db9abd-d34a-4628-9073-f76c5ee4ea7a" (UID: "e6db9abd-d34a-4628-9073-f76c5ee4ea7a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.736275 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-config" (OuterVolumeSpecName: "config") pod "e6db9abd-d34a-4628-9073-f76c5ee4ea7a" (UID: "e6db9abd-d34a-4628-9073-f76c5ee4ea7a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.741145 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6db9abd-d34a-4628-9073-f76c5ee4ea7a" (UID: "e6db9abd-d34a-4628-9073-f76c5ee4ea7a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.804778 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.805058 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.805067 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.805076 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6db9abd-d34a-4628-9073-f76c5ee4ea7a-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.892466 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.907187 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb831c8-3f64-44d9-b867-be31f64d4388","Type":"ContainerStarted","Data":"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60"} Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.924011 4726 generic.go:334] "Generic (PLEG): container finished" podID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" containerID="a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5" exitCode=0 Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.924073 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" event={"ID":"e6db9abd-d34a-4628-9073-f76c5ee4ea7a","Type":"ContainerDied","Data":"a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5"} Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.924103 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" event={"ID":"e6db9abd-d34a-4628-9073-f76c5ee4ea7a","Type":"ContainerDied","Data":"e0590802f295cf0f163ebc6032b626cb7c28ca28326c50ce24431a44cb1fb853"} Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.924121 4726 scope.go:117] "RemoveContainer" containerID="a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.924239 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-844b557b9c-r52hj" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.937125 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67dea71d-9dd9-4945-812e-df9a679aa505","Type":"ContainerStarted","Data":"06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690"} Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.959534 4726 scope.go:117] "RemoveContainer" containerID="c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.970784 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-r52hj"] Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.976002 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-ppzwh" event={"ID":"b5b50700-1688-4d3f-8508-30a069c33ac7","Type":"ContainerStarted","Data":"db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8"} Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.976061 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:41 crc kubenswrapper[4726]: I1123 20:25:41.991202 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-844b557b9c-r52hj"] Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.014904 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-775457b975-ppzwh" podStartSLOduration=3.014885033 podStartE2EDuration="3.014885033s" podCreationTimestamp="2025-11-23 20:25:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:42.003217062 +0000 UTC m=+1050.152258018" watchObservedRunningTime="2025-11-23 20:25:42.014885033 +0000 UTC m=+1050.163925989" Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.027093 4726 scope.go:117] "RemoveContainer" containerID="a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5" Nov 23 20:25:42 crc kubenswrapper[4726]: E1123 20:25:42.032153 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5\": container with ID starting with a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5 not found: ID does not exist" containerID="a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5" Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.032191 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5"} err="failed to get container status \"a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5\": rpc error: code = NotFound desc = could not find container \"a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5\": container with ID starting with a9238f6fdceb42a76164e70ac6a21dd0ab9763f29cd3de89fbbfc2c093dee7e5 not found: ID does not exist" Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.032211 4726 scope.go:117] "RemoveContainer" containerID="c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53" Nov 23 20:25:42 crc kubenswrapper[4726]: E1123 20:25:42.032703 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53\": container with ID starting with c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53 not found: ID does not exist" containerID="c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53" Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.032723 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53"} err="failed to get container status \"c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53\": rpc error: code = NotFound desc = could not find container \"c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53\": container with ID starting with c03ad8ca8689859dc965587d4bdbf91045fe64b246b2d77f09d34f644b62ba53 not found: ID does not exist" Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.599158 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" path="/var/lib/kubelet/pods/e6db9abd-d34a-4628-9073-f76c5ee4ea7a/volumes" Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.991571 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb831c8-3f64-44d9-b867-be31f64d4388","Type":"ContainerStarted","Data":"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67"} Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.991858 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerName="cinder-api-log" containerID="cri-o://2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60" gracePeriod=30 Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.992151 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 23 20:25:42 crc kubenswrapper[4726]: I1123 20:25:42.992311 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerName="cinder-api" containerID="cri-o://62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67" gracePeriod=30 Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.000802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67dea71d-9dd9-4945-812e-df9a679aa505","Type":"ContainerStarted","Data":"6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b"} Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.022695 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.022650518 podStartE2EDuration="4.022650518s" podCreationTimestamp="2025-11-23 20:25:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:43.017619984 +0000 UTC m=+1051.166660940" watchObservedRunningTime="2025-11-23 20:25:43.022650518 +0000 UTC m=+1051.171691484" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.044658 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.436384325 podStartE2EDuration="4.044450579s" podCreationTimestamp="2025-11-23 20:25:39 +0000 UTC" firstStartedPulling="2025-11-23 20:25:40.050756542 +0000 UTC m=+1048.199797498" lastFinishedPulling="2025-11-23 20:25:40.658822796 +0000 UTC m=+1048.807863752" observedRunningTime="2025-11-23 20:25:43.038030228 +0000 UTC m=+1051.187071184" watchObservedRunningTime="2025-11-23 20:25:43.044450579 +0000 UTC m=+1051.193491545" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.628540 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757071 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-scripts\") pod \"8cb831c8-3f64-44d9-b867-be31f64d4388\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757161 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnlxv\" (UniqueName: \"kubernetes.io/projected/8cb831c8-3f64-44d9-b867-be31f64d4388-kube-api-access-pnlxv\") pod \"8cb831c8-3f64-44d9-b867-be31f64d4388\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757186 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data\") pod \"8cb831c8-3f64-44d9-b867-be31f64d4388\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757206 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb831c8-3f64-44d9-b867-be31f64d4388-logs\") pod \"8cb831c8-3f64-44d9-b867-be31f64d4388\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757275 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data-custom\") pod \"8cb831c8-3f64-44d9-b867-be31f64d4388\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757317 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-combined-ca-bundle\") pod \"8cb831c8-3f64-44d9-b867-be31f64d4388\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757336 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb831c8-3f64-44d9-b867-be31f64d4388-etc-machine-id\") pod \"8cb831c8-3f64-44d9-b867-be31f64d4388\" (UID: \"8cb831c8-3f64-44d9-b867-be31f64d4388\") " Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757734 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cb831c8-3f64-44d9-b867-be31f64d4388-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8cb831c8-3f64-44d9-b867-be31f64d4388" (UID: "8cb831c8-3f64-44d9-b867-be31f64d4388"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.757774 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cb831c8-3f64-44d9-b867-be31f64d4388-logs" (OuterVolumeSpecName: "logs") pod "8cb831c8-3f64-44d9-b867-be31f64d4388" (UID: "8cb831c8-3f64-44d9-b867-be31f64d4388"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.767138 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cb831c8-3f64-44d9-b867-be31f64d4388-kube-api-access-pnlxv" (OuterVolumeSpecName: "kube-api-access-pnlxv") pod "8cb831c8-3f64-44d9-b867-be31f64d4388" (UID: "8cb831c8-3f64-44d9-b867-be31f64d4388"). InnerVolumeSpecName "kube-api-access-pnlxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.771502 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8cb831c8-3f64-44d9-b867-be31f64d4388" (UID: "8cb831c8-3f64-44d9-b867-be31f64d4388"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.777145 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-scripts" (OuterVolumeSpecName: "scripts") pod "8cb831c8-3f64-44d9-b867-be31f64d4388" (UID: "8cb831c8-3f64-44d9-b867-be31f64d4388"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.800013 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cb831c8-3f64-44d9-b867-be31f64d4388" (UID: "8cb831c8-3f64-44d9-b867-be31f64d4388"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.833353 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data" (OuterVolumeSpecName: "config-data") pod "8cb831c8-3f64-44d9-b867-be31f64d4388" (UID: "8cb831c8-3f64-44d9-b867-be31f64d4388"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.859809 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.859836 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.859847 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb831c8-3f64-44d9-b867-be31f64d4388-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.859858 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.859883 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnlxv\" (UniqueName: \"kubernetes.io/projected/8cb831c8-3f64-44d9-b867-be31f64d4388-kube-api-access-pnlxv\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.859896 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb831c8-3f64-44d9-b867-be31f64d4388-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:43 crc kubenswrapper[4726]: I1123 20:25:43.859905 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb831c8-3f64-44d9-b867-be31f64d4388-logs\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.012275 4726 generic.go:334] "Generic (PLEG): container finished" podID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerID="62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67" exitCode=0 Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.012304 4726 generic.go:334] "Generic (PLEG): container finished" podID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerID="2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60" exitCode=143 Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.012581 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.013181 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb831c8-3f64-44d9-b867-be31f64d4388","Type":"ContainerDied","Data":"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67"} Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.013208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb831c8-3f64-44d9-b867-be31f64d4388","Type":"ContainerDied","Data":"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60"} Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.013218 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb831c8-3f64-44d9-b867-be31f64d4388","Type":"ContainerDied","Data":"c74f09073460d9b3b88acd7ff45817a3e462027cb2b4b2d9547279c4c9a892af"} Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.013232 4726 scope.go:117] "RemoveContainer" containerID="62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.060937 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.088951 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.095706 4726 scope.go:117] "RemoveContainer" containerID="2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.118445 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:44 crc kubenswrapper[4726]: E1123 20:25:44.119127 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" containerName="dnsmasq-dns" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.119213 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" containerName="dnsmasq-dns" Nov 23 20:25:44 crc kubenswrapper[4726]: E1123 20:25:44.119333 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerName="cinder-api-log" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.119388 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerName="cinder-api-log" Nov 23 20:25:44 crc kubenswrapper[4726]: E1123 20:25:44.119466 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerName="cinder-api" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.119523 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerName="cinder-api" Nov 23 20:25:44 crc kubenswrapper[4726]: E1123 20:25:44.119581 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" containerName="init" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.119655 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" containerName="init" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.119950 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerName="cinder-api-log" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.120033 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6db9abd-d34a-4628-9073-f76c5ee4ea7a" containerName="dnsmasq-dns" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.120087 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" containerName="cinder-api" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.121097 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.123610 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.124496 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.125137 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.128544 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.165989 4726 scope.go:117] "RemoveContainer" containerID="62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67" Nov 23 20:25:44 crc kubenswrapper[4726]: E1123 20:25:44.167369 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67\": container with ID starting with 62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67 not found: ID does not exist" containerID="62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.167403 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67"} err="failed to get container status \"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67\": rpc error: code = NotFound desc = could not find container \"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67\": container with ID starting with 62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67 not found: ID does not exist" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.167426 4726 scope.go:117] "RemoveContainer" containerID="2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60" Nov 23 20:25:44 crc kubenswrapper[4726]: E1123 20:25:44.171013 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60\": container with ID starting with 2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60 not found: ID does not exist" containerID="2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.171056 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60"} err="failed to get container status \"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60\": rpc error: code = NotFound desc = could not find container \"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60\": container with ID starting with 2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60 not found: ID does not exist" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.171083 4726 scope.go:117] "RemoveContainer" containerID="62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.171739 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67"} err="failed to get container status \"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67\": rpc error: code = NotFound desc = could not find container \"62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67\": container with ID starting with 62deaa61422337dc313a4bb0a487bae513321c334bcd1a301b4d90140635ef67 not found: ID does not exist" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.171767 4726 scope.go:117] "RemoveContainer" containerID="2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.172937 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60"} err="failed to get container status \"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60\": rpc error: code = NotFound desc = could not find container \"2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60\": container with ID starting with 2c5dd21716c1bff18eeedf473493879d71e3b9109479d883d2d1e33bc8133e60 not found: ID does not exist" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273574 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzdzj\" (UniqueName: \"kubernetes.io/projected/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-kube-api-access-hzdzj\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273617 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-scripts\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273662 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273685 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-config-data-custom\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273717 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-logs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273745 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273765 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273792 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.273811 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-config-data\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375765 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-config-data-custom\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375801 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-logs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375832 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375859 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375903 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375924 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-config-data\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375979 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzdzj\" (UniqueName: \"kubernetes.io/projected/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-kube-api-access-hzdzj\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.375997 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-scripts\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.376423 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.376592 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-logs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.383844 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-config-data\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.386068 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.390378 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.393448 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-scripts\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.395347 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.395920 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-config-data-custom\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.397805 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzdzj\" (UniqueName: \"kubernetes.io/projected/3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275-kube-api-access-hzdzj\") pod \"cinder-api-0\" (UID: \"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275\") " pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.411250 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.472492 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.619921 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cb831c8-3f64-44d9-b867-be31f64d4388" path="/var/lib/kubelet/pods/8cb831c8-3f64-44d9-b867-be31f64d4388/volumes" Nov 23 20:25:44 crc kubenswrapper[4726]: I1123 20:25:44.990486 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 23 20:25:45 crc kubenswrapper[4726]: I1123 20:25:45.025478 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275","Type":"ContainerStarted","Data":"c0e325e9f5aa6d348b909e8482f63bf643bd1c82cfbc36f37183bcb3340fb31e"} Nov 23 20:25:46 crc kubenswrapper[4726]: I1123 20:25:46.042702 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275","Type":"ContainerStarted","Data":"dd82fc01d68d3825049082cd1a2b1073aa30489d560c64a81f84cd2b6c676934"} Nov 23 20:25:46 crc kubenswrapper[4726]: I1123 20:25:46.491411 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:46 crc kubenswrapper[4726]: I1123 20:25:46.552594 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:47 crc kubenswrapper[4726]: I1123 20:25:47.053721 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275","Type":"ContainerStarted","Data":"8054a11837b8d53e6abe942df05df07787593b6583c16d8c54f2087159e73517"} Nov 23 20:25:47 crc kubenswrapper[4726]: I1123 20:25:47.075608 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.0755890790000002 podStartE2EDuration="3.075589079s" podCreationTimestamp="2025-11-23 20:25:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:47.069148337 +0000 UTC m=+1055.218189293" watchObservedRunningTime="2025-11-23 20:25:47.075589079 +0000 UTC m=+1055.224630025" Nov 23 20:25:48 crc kubenswrapper[4726]: I1123 20:25:48.061620 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.359608 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.541543 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85855866bd-77958" Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.611064 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.614062 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-595b84cd88-8zlj4"] Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.614314 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-595b84cd88-8zlj4" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api-log" containerID="cri-o://4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33" gracePeriod=30 Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.614466 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-595b84cd88-8zlj4" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api" containerID="cri-o://e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3" gracePeriod=30 Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.700259 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-skj9z"] Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.700800 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" podUID="80352774-1f1a-4e54-b4e2-2fad013814e8" containerName="dnsmasq-dns" containerID="cri-o://3f37211739e8493c4df510d6b3dba50a595edf0fefd4ac39c053103a173bf29d" gracePeriod=10 Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.776493 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 23 20:25:49 crc kubenswrapper[4726]: I1123 20:25:49.837289 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.080818 4726 generic.go:334] "Generic (PLEG): container finished" podID="80352774-1f1a-4e54-b4e2-2fad013814e8" containerID="3f37211739e8493c4df510d6b3dba50a595edf0fefd4ac39c053103a173bf29d" exitCode=0 Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.080902 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" event={"ID":"80352774-1f1a-4e54-b4e2-2fad013814e8","Type":"ContainerDied","Data":"3f37211739e8493c4df510d6b3dba50a595edf0fefd4ac39c053103a173bf29d"} Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.083546 4726 generic.go:334] "Generic (PLEG): container finished" podID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerID="4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33" exitCode=143 Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.083629 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-595b84cd88-8zlj4" event={"ID":"9a95fa41-6d0c-44e2-8e87-2045ef8bc345","Type":"ContainerDied","Data":"4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33"} Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.083798 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" containerName="probe" containerID="cri-o://6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b" gracePeriod=30 Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.083946 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" containerName="cinder-scheduler" containerID="cri-o://06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690" gracePeriod=30 Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.258131 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.391795 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-nb\") pod \"80352774-1f1a-4e54-b4e2-2fad013814e8\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.391933 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-dns-svc\") pod \"80352774-1f1a-4e54-b4e2-2fad013814e8\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.391992 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-config\") pod \"80352774-1f1a-4e54-b4e2-2fad013814e8\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.392010 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rjjj\" (UniqueName: \"kubernetes.io/projected/80352774-1f1a-4e54-b4e2-2fad013814e8-kube-api-access-2rjjj\") pod \"80352774-1f1a-4e54-b4e2-2fad013814e8\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.392090 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-sb\") pod \"80352774-1f1a-4e54-b4e2-2fad013814e8\" (UID: \"80352774-1f1a-4e54-b4e2-2fad013814e8\") " Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.426800 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80352774-1f1a-4e54-b4e2-2fad013814e8-kube-api-access-2rjjj" (OuterVolumeSpecName: "kube-api-access-2rjjj") pod "80352774-1f1a-4e54-b4e2-2fad013814e8" (UID: "80352774-1f1a-4e54-b4e2-2fad013814e8"). InnerVolumeSpecName "kube-api-access-2rjjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.457955 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "80352774-1f1a-4e54-b4e2-2fad013814e8" (UID: "80352774-1f1a-4e54-b4e2-2fad013814e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.467333 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-config" (OuterVolumeSpecName: "config") pod "80352774-1f1a-4e54-b4e2-2fad013814e8" (UID: "80352774-1f1a-4e54-b4e2-2fad013814e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.483351 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "80352774-1f1a-4e54-b4e2-2fad013814e8" (UID: "80352774-1f1a-4e54-b4e2-2fad013814e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.485501 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "80352774-1f1a-4e54-b4e2-2fad013814e8" (UID: "80352774-1f1a-4e54-b4e2-2fad013814e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.494223 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.494250 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.494260 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.494270 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rjjj\" (UniqueName: \"kubernetes.io/projected/80352774-1f1a-4e54-b4e2-2fad013814e8-kube-api-access-2rjjj\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.494281 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80352774-1f1a-4e54-b4e2-2fad013814e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:50 crc kubenswrapper[4726]: I1123 20:25:50.822194 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 23 20:25:51 crc kubenswrapper[4726]: I1123 20:25:51.092595 4726 generic.go:334] "Generic (PLEG): container finished" podID="67dea71d-9dd9-4945-812e-df9a679aa505" containerID="6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b" exitCode=0 Nov 23 20:25:51 crc kubenswrapper[4726]: I1123 20:25:51.092657 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67dea71d-9dd9-4945-812e-df9a679aa505","Type":"ContainerDied","Data":"6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b"} Nov 23 20:25:51 crc kubenswrapper[4726]: I1123 20:25:51.095078 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" event={"ID":"80352774-1f1a-4e54-b4e2-2fad013814e8","Type":"ContainerDied","Data":"d0471f4b6bcd416320fa8e86289fc560a80ce74ba68d8e4b35c5ae4b8ad09942"} Nov 23 20:25:51 crc kubenswrapper[4726]: I1123 20:25:51.095123 4726 scope.go:117] "RemoveContainer" containerID="3f37211739e8493c4df510d6b3dba50a595edf0fefd4ac39c053103a173bf29d" Nov 23 20:25:51 crc kubenswrapper[4726]: I1123 20:25:51.095131 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66f4bdbdb7-skj9z" Nov 23 20:25:51 crc kubenswrapper[4726]: I1123 20:25:51.123957 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-skj9z"] Nov 23 20:25:51 crc kubenswrapper[4726]: I1123 20:25:51.138040 4726 scope.go:117] "RemoveContainer" containerID="7748370b657c9f474127749f38bba264a2fe20502c2cd6e0cb805cd13904285d" Nov 23 20:25:51 crc kubenswrapper[4726]: I1123 20:25:51.143565 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66f4bdbdb7-skj9z"] Nov 23 20:25:52 crc kubenswrapper[4726]: I1123 20:25:52.599882 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80352774-1f1a-4e54-b4e2-2fad013814e8" path="/var/lib/kubelet/pods/80352774-1f1a-4e54-b4e2-2fad013814e8/volumes" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.050847 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-595b84cd88-8zlj4" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": read tcp 10.217.0.2:36066->10.217.0.148:9311: read: connection reset by peer" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.051742 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-595b84cd88-8zlj4" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": read tcp 10.217.0.2:36078->10.217.0.148:9311: read: connection reset by peer" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.512806 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.651486 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data-custom\") pod \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.651524 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtv5p\" (UniqueName: \"kubernetes.io/projected/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-kube-api-access-jtv5p\") pod \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.651591 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data\") pod \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.651645 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-combined-ca-bundle\") pod \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.651724 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-logs\") pod \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\" (UID: \"9a95fa41-6d0c-44e2-8e87-2045ef8bc345\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.652662 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-logs" (OuterVolumeSpecName: "logs") pod "9a95fa41-6d0c-44e2-8e87-2045ef8bc345" (UID: "9a95fa41-6d0c-44e2-8e87-2045ef8bc345"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.652946 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-logs\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.672640 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9a95fa41-6d0c-44e2-8e87-2045ef8bc345" (UID: "9a95fa41-6d0c-44e2-8e87-2045ef8bc345"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.673154 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-kube-api-access-jtv5p" (OuterVolumeSpecName: "kube-api-access-jtv5p") pod "9a95fa41-6d0c-44e2-8e87-2045ef8bc345" (UID: "9a95fa41-6d0c-44e2-8e87-2045ef8bc345"). InnerVolumeSpecName "kube-api-access-jtv5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.677998 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a95fa41-6d0c-44e2-8e87-2045ef8bc345" (UID: "9a95fa41-6d0c-44e2-8e87-2045ef8bc345"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.701643 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data" (OuterVolumeSpecName: "config-data") pod "9a95fa41-6d0c-44e2-8e87-2045ef8bc345" (UID: "9a95fa41-6d0c-44e2-8e87-2045ef8bc345"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.754844 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.755095 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtv5p\" (UniqueName: \"kubernetes.io/projected/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-kube-api-access-jtv5p\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.755107 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.755115 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a95fa41-6d0c-44e2-8e87-2045ef8bc345-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.912762 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.958693 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67dea71d-9dd9-4945-812e-df9a679aa505-etc-machine-id\") pod \"67dea71d-9dd9-4945-812e-df9a679aa505\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.958783 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-scripts\") pod \"67dea71d-9dd9-4945-812e-df9a679aa505\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.958821 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data\") pod \"67dea71d-9dd9-4945-812e-df9a679aa505\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.959164 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-combined-ca-bundle\") pod \"67dea71d-9dd9-4945-812e-df9a679aa505\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.959205 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pskwm\" (UniqueName: \"kubernetes.io/projected/67dea71d-9dd9-4945-812e-df9a679aa505-kube-api-access-pskwm\") pod \"67dea71d-9dd9-4945-812e-df9a679aa505\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.959253 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data-custom\") pod \"67dea71d-9dd9-4945-812e-df9a679aa505\" (UID: \"67dea71d-9dd9-4945-812e-df9a679aa505\") " Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.960258 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67dea71d-9dd9-4945-812e-df9a679aa505-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "67dea71d-9dd9-4945-812e-df9a679aa505" (UID: "67dea71d-9dd9-4945-812e-df9a679aa505"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.964160 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-scripts" (OuterVolumeSpecName: "scripts") pod "67dea71d-9dd9-4945-812e-df9a679aa505" (UID: "67dea71d-9dd9-4945-812e-df9a679aa505"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.966843 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67dea71d-9dd9-4945-812e-df9a679aa505-kube-api-access-pskwm" (OuterVolumeSpecName: "kube-api-access-pskwm") pod "67dea71d-9dd9-4945-812e-df9a679aa505" (UID: "67dea71d-9dd9-4945-812e-df9a679aa505"). InnerVolumeSpecName "kube-api-access-pskwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:25:53 crc kubenswrapper[4726]: I1123 20:25:53.967030 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "67dea71d-9dd9-4945-812e-df9a679aa505" (UID: "67dea71d-9dd9-4945-812e-df9a679aa505"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.031992 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67dea71d-9dd9-4945-812e-df9a679aa505" (UID: "67dea71d-9dd9-4945-812e-df9a679aa505"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.060910 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67dea71d-9dd9-4945-812e-df9a679aa505-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.060945 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.060955 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.060965 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pskwm\" (UniqueName: \"kubernetes.io/projected/67dea71d-9dd9-4945-812e-df9a679aa505-kube-api-access-pskwm\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.060975 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.064011 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data" (OuterVolumeSpecName: "config-data") pod "67dea71d-9dd9-4945-812e-df9a679aa505" (UID: "67dea71d-9dd9-4945-812e-df9a679aa505"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.128462 4726 generic.go:334] "Generic (PLEG): container finished" podID="67dea71d-9dd9-4945-812e-df9a679aa505" containerID="06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690" exitCode=0 Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.128523 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67dea71d-9dd9-4945-812e-df9a679aa505","Type":"ContainerDied","Data":"06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690"} Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.128553 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"67dea71d-9dd9-4945-812e-df9a679aa505","Type":"ContainerDied","Data":"49179fca25bde95843efa7ecee03089e80e13529f9b5a842c3f7295a861be3a2"} Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.128569 4726 scope.go:117] "RemoveContainer" containerID="6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.128721 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.140723 4726 generic.go:334] "Generic (PLEG): container finished" podID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerID="e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3" exitCode=0 Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.140790 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-595b84cd88-8zlj4" event={"ID":"9a95fa41-6d0c-44e2-8e87-2045ef8bc345","Type":"ContainerDied","Data":"e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3"} Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.140815 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-595b84cd88-8zlj4" event={"ID":"9a95fa41-6d0c-44e2-8e87-2045ef8bc345","Type":"ContainerDied","Data":"7c32aa99bc85a3f27774f3d6a189973f74993a62cf08bbc91716ccc5e3e62606"} Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.140826 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-595b84cd88-8zlj4" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.164315 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67dea71d-9dd9-4945-812e-df9a679aa505-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.196092 4726 scope.go:117] "RemoveContainer" containerID="06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.201025 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.209307 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.223713 4726 scope.go:117] "RemoveContainer" containerID="6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.224332 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b\": container with ID starting with 6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b not found: ID does not exist" containerID="6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.224383 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b"} err="failed to get container status \"6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b\": rpc error: code = NotFound desc = could not find container \"6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b\": container with ID starting with 6ea126e258e77ca937c6182ed6d0cdf766707e5e2c9aa17dce042bf3a48ff19b not found: ID does not exist" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.224551 4726 scope.go:117] "RemoveContainer" containerID="06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.224853 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690\": container with ID starting with 06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690 not found: ID does not exist" containerID="06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.224911 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690"} err="failed to get container status \"06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690\": rpc error: code = NotFound desc = could not find container \"06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690\": container with ID starting with 06c9702f26bb7d285ca5bb873f34d65c82e526bd4297e9914768e59709e97690 not found: ID does not exist" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.224939 4726 scope.go:117] "RemoveContainer" containerID="e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.225753 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-595b84cd88-8zlj4"] Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.235582 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-595b84cd88-8zlj4"] Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.243206 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.244062 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" containerName="cinder-scheduler" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.244188 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" containerName="cinder-scheduler" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.244271 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api-log" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.244332 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api-log" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.244407 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.244485 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.244544 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" containerName="probe" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.244600 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" containerName="probe" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.244652 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80352774-1f1a-4e54-b4e2-2fad013814e8" containerName="init" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.244712 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="80352774-1f1a-4e54-b4e2-2fad013814e8" containerName="init" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.244803 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80352774-1f1a-4e54-b4e2-2fad013814e8" containerName="dnsmasq-dns" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.244863 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="80352774-1f1a-4e54-b4e2-2fad013814e8" containerName="dnsmasq-dns" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.245086 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" containerName="probe" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.245153 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" containerName="cinder-scheduler" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.245215 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.246046 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" containerName="barbican-api-log" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.246152 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="80352774-1f1a-4e54-b4e2-2fad013814e8" containerName="dnsmasq-dns" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.247377 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.248803 4726 scope.go:117] "RemoveContainer" containerID="4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.251801 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.261215 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.275321 4726 scope.go:117] "RemoveContainer" containerID="e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.278997 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3\": container with ID starting with e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3 not found: ID does not exist" containerID="e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.279040 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3"} err="failed to get container status \"e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3\": rpc error: code = NotFound desc = could not find container \"e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3\": container with ID starting with e834b47c587dececd6605d0bfaff8cee0b6299ab20f47248de1a8e743db6f4d3 not found: ID does not exist" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.279069 4726 scope.go:117] "RemoveContainer" containerID="4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33" Nov 23 20:25:54 crc kubenswrapper[4726]: E1123 20:25:54.282033 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33\": container with ID starting with 4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33 not found: ID does not exist" containerID="4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.282154 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33"} err="failed to get container status \"4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33\": rpc error: code = NotFound desc = could not find container \"4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33\": container with ID starting with 4252f8eddb2bfb085bf7fdafefa8662375e5ab1cc48477176b80f26244e1db33 not found: ID does not exist" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.376649 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.376993 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.377220 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-scripts\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.377301 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.377377 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2492f\" (UniqueName: \"kubernetes.io/projected/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-kube-api-access-2492f\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.377453 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-config-data\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.478763 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.478807 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2492f\" (UniqueName: \"kubernetes.io/projected/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-kube-api-access-2492f\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.478840 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-config-data\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.478878 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.478905 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.479014 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-scripts\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.479324 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.484454 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-scripts\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.485833 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.487510 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-config-data\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.497506 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2492f\" (UniqueName: \"kubernetes.io/projected/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-kube-api-access-2492f\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.503558 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5462b4dd-3aaf-4511-b08f-4cfe23a07a0d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d\") " pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.583757 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.598984 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67dea71d-9dd9-4945-812e-df9a679aa505" path="/var/lib/kubelet/pods/67dea71d-9dd9-4945-812e-df9a679aa505/volumes" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.599849 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a95fa41-6d0c-44e2-8e87-2045ef8bc345" path="/var/lib/kubelet/pods/9a95fa41-6d0c-44e2-8e87-2045ef8bc345/volumes" Nov 23 20:25:54 crc kubenswrapper[4726]: I1123 20:25:54.856138 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:25:55 crc kubenswrapper[4726]: I1123 20:25:55.118247 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 23 20:25:55 crc kubenswrapper[4726]: I1123 20:25:55.153315 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d","Type":"ContainerStarted","Data":"37133b8750e2b1a5da48201c5035c6647457e3e708b7b4a1b240a00abb71ee42"} Nov 23 20:25:56 crc kubenswrapper[4726]: I1123 20:25:56.057701 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:56 crc kubenswrapper[4726]: I1123 20:25:56.065381 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-fbb56fbb6-7fzgk" Nov 23 20:25:56 crc kubenswrapper[4726]: I1123 20:25:56.192245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d","Type":"ContainerStarted","Data":"370f108254833dc3d1bdbeeaefa9402f8597cd89df6cdf798b064b80145bfc19"} Nov 23 20:25:56 crc kubenswrapper[4726]: I1123 20:25:56.692258 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-b76988558-2489x" Nov 23 20:25:57 crc kubenswrapper[4726]: I1123 20:25:57.047411 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 23 20:25:57 crc kubenswrapper[4726]: I1123 20:25:57.204778 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5462b4dd-3aaf-4511-b08f-4cfe23a07a0d","Type":"ContainerStarted","Data":"c6ab896168f6aa5a05a2d29879ef66c502acb614183649520eed771c23014fb3"} Nov 23 20:25:57 crc kubenswrapper[4726]: I1123 20:25:57.222353 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.222337109 podStartE2EDuration="3.222337109s" podCreationTimestamp="2025-11-23 20:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:25:57.221837755 +0000 UTC m=+1065.370878721" watchObservedRunningTime="2025-11-23 20:25:57.222337109 +0000 UTC m=+1065.371378055" Nov 23 20:25:57 crc kubenswrapper[4726]: I1123 20:25:57.249586 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-57cf6d944c-r5qgs" Nov 23 20:25:57 crc kubenswrapper[4726]: I1123 20:25:57.356292 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c5d686b4-gshjj"] Nov 23 20:25:57 crc kubenswrapper[4726]: I1123 20:25:57.356568 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c5d686b4-gshjj" podUID="34abb904-d204-4941-afc7-21887ef6703c" containerName="neutron-api" containerID="cri-o://f375e027d360f6079f3519e5b5529c65a935a08a49b4be2f20daaa9e461a04d6" gracePeriod=30 Nov 23 20:25:57 crc kubenswrapper[4726]: I1123 20:25:57.356722 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c5d686b4-gshjj" podUID="34abb904-d204-4941-afc7-21887ef6703c" containerName="neutron-httpd" containerID="cri-o://5ba3211414beac95a6a7fbc5f238d09f393b54bf60c0601a3676dda0449a0a62" gracePeriod=30 Nov 23 20:25:58 crc kubenswrapper[4726]: I1123 20:25:58.214376 4726 generic.go:334] "Generic (PLEG): container finished" podID="34abb904-d204-4941-afc7-21887ef6703c" containerID="5ba3211414beac95a6a7fbc5f238d09f393b54bf60c0601a3676dda0449a0a62" exitCode=0 Nov 23 20:25:58 crc kubenswrapper[4726]: I1123 20:25:58.214466 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c5d686b4-gshjj" event={"ID":"34abb904-d204-4941-afc7-21887ef6703c","Type":"ContainerDied","Data":"5ba3211414beac95a6a7fbc5f238d09f393b54bf60c0601a3676dda0449a0a62"} Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.206089 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.207972 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.210996 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-22g86" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.214146 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.214361 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.233240 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.268321 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-combined-ca-bundle\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.269075 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.269299 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config-secret\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.269573 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krlxv\" (UniqueName: \"kubernetes.io/projected/db96f9c1-a887-4701-8a13-030e50364efe-kube-api-access-krlxv\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.370781 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.370859 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config-secret\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.370954 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krlxv\" (UniqueName: \"kubernetes.io/projected/db96f9c1-a887-4701-8a13-030e50364efe-kube-api-access-krlxv\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.370986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-combined-ca-bundle\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.372043 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.377062 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config-secret\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.379280 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-combined-ca-bundle\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.397467 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krlxv\" (UniqueName: \"kubernetes.io/projected/db96f9c1-a887-4701-8a13-030e50364efe-kube-api-access-krlxv\") pod \"openstackclient\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.535837 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.584470 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.652916 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.687170 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.710954 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.712182 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.738148 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.784363 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-openstack-config-secret\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.784483 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.784511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-openstack-config\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.784553 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvcn2\" (UniqueName: \"kubernetes.io/projected/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-kube-api-access-fvcn2\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: E1123 20:25:59.805318 4726 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 23 20:25:59 crc kubenswrapper[4726]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_db96f9c1-a887-4701-8a13-030e50364efe_0(ac4e31c6bd49155dc03db23c85e9c28030ebc05740decce8163ce330465740ff): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ac4e31c6bd49155dc03db23c85e9c28030ebc05740decce8163ce330465740ff" Netns:"/var/run/netns/2e1a820b-4f34-485d-9383-8baa5df134a8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=ac4e31c6bd49155dc03db23c85e9c28030ebc05740decce8163ce330465740ff;K8S_POD_UID=db96f9c1-a887-4701-8a13-030e50364efe" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/db96f9c1-a887-4701-8a13-030e50364efe]: expected pod UID "db96f9c1-a887-4701-8a13-030e50364efe" but got "bfa50ecd-0a29-44c8-b219-d2e5df3709d1" from Kube API Nov 23 20:25:59 crc kubenswrapper[4726]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 23 20:25:59 crc kubenswrapper[4726]: > Nov 23 20:25:59 crc kubenswrapper[4726]: E1123 20:25:59.805393 4726 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 23 20:25:59 crc kubenswrapper[4726]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_db96f9c1-a887-4701-8a13-030e50364efe_0(ac4e31c6bd49155dc03db23c85e9c28030ebc05740decce8163ce330465740ff): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ac4e31c6bd49155dc03db23c85e9c28030ebc05740decce8163ce330465740ff" Netns:"/var/run/netns/2e1a820b-4f34-485d-9383-8baa5df134a8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=ac4e31c6bd49155dc03db23c85e9c28030ebc05740decce8163ce330465740ff;K8S_POD_UID=db96f9c1-a887-4701-8a13-030e50364efe" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/db96f9c1-a887-4701-8a13-030e50364efe]: expected pod UID "db96f9c1-a887-4701-8a13-030e50364efe" but got "bfa50ecd-0a29-44c8-b219-d2e5df3709d1" from Kube API Nov 23 20:25:59 crc kubenswrapper[4726]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 23 20:25:59 crc kubenswrapper[4726]: > pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.886332 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvcn2\" (UniqueName: \"kubernetes.io/projected/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-kube-api-access-fvcn2\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.886428 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-openstack-config-secret\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.886510 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.886540 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-openstack-config\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.887382 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-openstack-config\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.891121 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.891143 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-openstack-config-secret\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:25:59 crc kubenswrapper[4726]: I1123 20:25:59.904518 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvcn2\" (UniqueName: \"kubernetes.io/projected/bfa50ecd-0a29-44c8-b219-d2e5df3709d1-kube-api-access-fvcn2\") pod \"openstackclient\" (UID: \"bfa50ecd-0a29-44c8-b219-d2e5df3709d1\") " pod="openstack/openstackclient" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.040936 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.266908 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.274265 4726 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="db96f9c1-a887-4701-8a13-030e50364efe" podUID="bfa50ecd-0a29-44c8-b219-d2e5df3709d1" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.301487 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.398626 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config-secret\") pod \"db96f9c1-a887-4701-8a13-030e50364efe\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.398774 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config\") pod \"db96f9c1-a887-4701-8a13-030e50364efe\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.398839 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-combined-ca-bundle\") pod \"db96f9c1-a887-4701-8a13-030e50364efe\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.398927 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krlxv\" (UniqueName: \"kubernetes.io/projected/db96f9c1-a887-4701-8a13-030e50364efe-kube-api-access-krlxv\") pod \"db96f9c1-a887-4701-8a13-030e50364efe\" (UID: \"db96f9c1-a887-4701-8a13-030e50364efe\") " Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.400745 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "db96f9c1-a887-4701-8a13-030e50364efe" (UID: "db96f9c1-a887-4701-8a13-030e50364efe"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.406834 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "db96f9c1-a887-4701-8a13-030e50364efe" (UID: "db96f9c1-a887-4701-8a13-030e50364efe"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.408523 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db96f9c1-a887-4701-8a13-030e50364efe" (UID: "db96f9c1-a887-4701-8a13-030e50364efe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.408621 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db96f9c1-a887-4701-8a13-030e50364efe-kube-api-access-krlxv" (OuterVolumeSpecName: "kube-api-access-krlxv") pod "db96f9c1-a887-4701-8a13-030e50364efe" (UID: "db96f9c1-a887-4701-8a13-030e50364efe"). InnerVolumeSpecName "kube-api-access-krlxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.501767 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krlxv\" (UniqueName: \"kubernetes.io/projected/db96f9c1-a887-4701-8a13-030e50364efe-kube-api-access-krlxv\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.501807 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.501818 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/db96f9c1-a887-4701-8a13-030e50364efe-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.501827 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db96f9c1-a887-4701-8a13-030e50364efe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.602970 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db96f9c1-a887-4701-8a13-030e50364efe" path="/var/lib/kubelet/pods/db96f9c1-a887-4701-8a13-030e50364efe/volumes" Nov 23 20:26:00 crc kubenswrapper[4726]: I1123 20:26:00.637385 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 23 20:26:01 crc kubenswrapper[4726]: I1123 20:26:01.281357 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 23 20:26:01 crc kubenswrapper[4726]: I1123 20:26:01.282568 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bfa50ecd-0a29-44c8-b219-d2e5df3709d1","Type":"ContainerStarted","Data":"d894f86827351ffdea4b86c56085f6a17104bea251e14d292688d67461ac06cd"} Nov 23 20:26:01 crc kubenswrapper[4726]: I1123 20:26:01.293288 4726 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="db96f9c1-a887-4701-8a13-030e50364efe" podUID="bfa50ecd-0a29-44c8-b219-d2e5df3709d1" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.314334 4726 generic.go:334] "Generic (PLEG): container finished" podID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerID="77c17bb00fe8203cb0bded7db33549111b5f33c679b58deefd29ae1d8baa3c32" exitCode=137 Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.315028 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerDied","Data":"77c17bb00fe8203cb0bded7db33549111b5f33c679b58deefd29ae1d8baa3c32"} Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.315098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582","Type":"ContainerDied","Data":"436aec86f2a0e4ef14ed59d529e9dbcde35f31327aafce2a73eb2efd91f90456"} Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.315136 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="436aec86f2a0e4ef14ed59d529e9dbcde35f31327aafce2a73eb2efd91f90456" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.366150 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.402172 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz6ht\" (UniqueName: \"kubernetes.io/projected/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-kube-api-access-rz6ht\") pod \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.402229 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-run-httpd\") pod \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.402557 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-combined-ca-bundle\") pod \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.402723 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-sg-core-conf-yaml\") pod \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.402835 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-config-data\") pod \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.402880 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-log-httpd\") pod \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.402947 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-scripts\") pod \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\" (UID: \"5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582\") " Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.404737 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" (UID: "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.415334 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" (UID: "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.440010 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" (UID: "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.486215 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-scripts" (OuterVolumeSpecName: "scripts") pod "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" (UID: "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.501821 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-kube-api-access-rz6ht" (OuterVolumeSpecName: "kube-api-access-rz6ht") pod "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" (UID: "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582"). InnerVolumeSpecName "kube-api-access-rz6ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.506718 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.506757 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.506768 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz6ht\" (UniqueName: \"kubernetes.io/projected/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-kube-api-access-rz6ht\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.506779 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.506789 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.561376 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-config-data" (OuterVolumeSpecName: "config-data") pod "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" (UID: "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.597165 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" (UID: "5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.608364 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.608390 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:04 crc kubenswrapper[4726]: I1123 20:26:04.934681 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.324510 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.354267 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.366177 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.398306 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:05 crc kubenswrapper[4726]: E1123 20:26:05.398783 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="ceilometer-central-agent" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.398813 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="ceilometer-central-agent" Nov 23 20:26:05 crc kubenswrapper[4726]: E1123 20:26:05.398839 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="proxy-httpd" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.398847 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="proxy-httpd" Nov 23 20:26:05 crc kubenswrapper[4726]: E1123 20:26:05.398887 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="ceilometer-notification-agent" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.398898 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="ceilometer-notification-agent" Nov 23 20:26:05 crc kubenswrapper[4726]: E1123 20:26:05.398912 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="sg-core" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.398920 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="sg-core" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.399145 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="ceilometer-notification-agent" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.399177 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="proxy-httpd" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.399205 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="ceilometer-central-agent" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.399236 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" containerName="sg-core" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.400848 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.403892 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.404001 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.423696 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.423741 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.423777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j26ht\" (UniqueName: \"kubernetes.io/projected/85ff1241-f8a5-4595-9ad5-2386fb6a7075-kube-api-access-j26ht\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.423801 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-run-httpd\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.423835 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-log-httpd\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.423922 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-scripts\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.423986 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-config-data\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.425835 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.527034 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.527103 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j26ht\" (UniqueName: \"kubernetes.io/projected/85ff1241-f8a5-4595-9ad5-2386fb6a7075-kube-api-access-j26ht\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.527147 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-run-httpd\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.527197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-log-httpd\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.527239 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-scripts\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.527335 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-config-data\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.528036 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.528080 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-run-httpd\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.528176 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-log-httpd\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.536564 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.545626 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-config-data\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.546309 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-scripts\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.547682 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j26ht\" (UniqueName: \"kubernetes.io/projected/85ff1241-f8a5-4595-9ad5-2386fb6a7075-kube-api-access-j26ht\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.559149 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.723648 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:05 crc kubenswrapper[4726]: I1123 20:26:05.872342 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:06 crc kubenswrapper[4726]: I1123 20:26:06.600267 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582" path="/var/lib/kubelet/pods/5e5b5d1b-5a67-46d9-aa6b-bdb0fed22582/volumes" Nov 23 20:26:09 crc kubenswrapper[4726]: I1123 20:26:09.043192 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:26:09 crc kubenswrapper[4726]: I1123 20:26:09.043516 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:26:11 crc kubenswrapper[4726]: I1123 20:26:11.379957 4726 generic.go:334] "Generic (PLEG): container finished" podID="34abb904-d204-4941-afc7-21887ef6703c" containerID="f375e027d360f6079f3519e5b5529c65a935a08a49b4be2f20daaa9e461a04d6" exitCode=0 Nov 23 20:26:11 crc kubenswrapper[4726]: I1123 20:26:11.380066 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c5d686b4-gshjj" event={"ID":"34abb904-d204-4941-afc7-21887ef6703c","Type":"ContainerDied","Data":"f375e027d360f6079f3519e5b5529c65a935a08a49b4be2f20daaa9e461a04d6"} Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.012572 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.057000 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rkls\" (UniqueName: \"kubernetes.io/projected/34abb904-d204-4941-afc7-21887ef6703c-kube-api-access-8rkls\") pod \"34abb904-d204-4941-afc7-21887ef6703c\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.057118 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-httpd-config\") pod \"34abb904-d204-4941-afc7-21887ef6703c\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.057262 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-config\") pod \"34abb904-d204-4941-afc7-21887ef6703c\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.057362 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-ovndb-tls-certs\") pod \"34abb904-d204-4941-afc7-21887ef6703c\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.057400 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-combined-ca-bundle\") pod \"34abb904-d204-4941-afc7-21887ef6703c\" (UID: \"34abb904-d204-4941-afc7-21887ef6703c\") " Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.063980 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34abb904-d204-4941-afc7-21887ef6703c-kube-api-access-8rkls" (OuterVolumeSpecName: "kube-api-access-8rkls") pod "34abb904-d204-4941-afc7-21887ef6703c" (UID: "34abb904-d204-4941-afc7-21887ef6703c"). InnerVolumeSpecName "kube-api-access-8rkls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.072973 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "34abb904-d204-4941-afc7-21887ef6703c" (UID: "34abb904-d204-4941-afc7-21887ef6703c"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.112045 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-config" (OuterVolumeSpecName: "config") pod "34abb904-d204-4941-afc7-21887ef6703c" (UID: "34abb904-d204-4941-afc7-21887ef6703c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.124938 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34abb904-d204-4941-afc7-21887ef6703c" (UID: "34abb904-d204-4941-afc7-21887ef6703c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.143569 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.157252 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "34abb904-d204-4941-afc7-21887ef6703c" (UID: "34abb904-d204-4941-afc7-21887ef6703c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.159433 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rkls\" (UniqueName: \"kubernetes.io/projected/34abb904-d204-4941-afc7-21887ef6703c-kube-api-access-8rkls\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.159472 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.159486 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.159496 4726 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.159506 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34abb904-d204-4941-afc7-21887ef6703c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.388321 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerStarted","Data":"d579eff835a65f2b6774995acb858dc20aa42cb88be9b7fe0f1b0a9bbc586be1"} Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.390321 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c5d686b4-gshjj" event={"ID":"34abb904-d204-4941-afc7-21887ef6703c","Type":"ContainerDied","Data":"c8ce54b856a9b1f7a597245b007458d38db72aada2788b847e730864b50c0f97"} Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.390353 4726 scope.go:117] "RemoveContainer" containerID="5ba3211414beac95a6a7fbc5f238d09f393b54bf60c0601a3676dda0449a0a62" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.390400 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c5d686b4-gshjj" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.395204 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bfa50ecd-0a29-44c8-b219-d2e5df3709d1","Type":"ContainerStarted","Data":"e74cf52d9e015158a5e06822a31dc53acee58910e86c3d0a3ab6069281217b03"} Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.411095 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.5233683300000003 podStartE2EDuration="13.411069114s" podCreationTimestamp="2025-11-23 20:25:59 +0000 UTC" firstStartedPulling="2025-11-23 20:26:00.665378308 +0000 UTC m=+1068.814419264" lastFinishedPulling="2025-11-23 20:26:11.553079092 +0000 UTC m=+1079.702120048" observedRunningTime="2025-11-23 20:26:12.408573778 +0000 UTC m=+1080.557614744" watchObservedRunningTime="2025-11-23 20:26:12.411069114 +0000 UTC m=+1080.560110070" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.414480 4726 scope.go:117] "RemoveContainer" containerID="f375e027d360f6079f3519e5b5529c65a935a08a49b4be2f20daaa9e461a04d6" Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.446249 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c5d686b4-gshjj"] Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.456942 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c5d686b4-gshjj"] Nov 23 20:26:12 crc kubenswrapper[4726]: I1123 20:26:12.619124 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34abb904-d204-4941-afc7-21887ef6703c" path="/var/lib/kubelet/pods/34abb904-d204-4941-afc7-21887ef6703c/volumes" Nov 23 20:26:13 crc kubenswrapper[4726]: I1123 20:26:13.410527 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerStarted","Data":"83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a"} Nov 23 20:26:14 crc kubenswrapper[4726]: I1123 20:26:14.421687 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerStarted","Data":"1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15"} Nov 23 20:26:14 crc kubenswrapper[4726]: I1123 20:26:14.422048 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerStarted","Data":"4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132"} Nov 23 20:26:17 crc kubenswrapper[4726]: I1123 20:26:17.453051 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerStarted","Data":"8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd"} Nov 23 20:26:17 crc kubenswrapper[4726]: I1123 20:26:17.453380 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="ceilometer-central-agent" containerID="cri-o://83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a" gracePeriod=30 Nov 23 20:26:17 crc kubenswrapper[4726]: I1123 20:26:17.453956 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="proxy-httpd" containerID="cri-o://8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd" gracePeriod=30 Nov 23 20:26:17 crc kubenswrapper[4726]: I1123 20:26:17.454035 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="sg-core" containerID="cri-o://1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15" gracePeriod=30 Nov 23 20:26:17 crc kubenswrapper[4726]: I1123 20:26:17.454112 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="ceilometer-notification-agent" containerID="cri-o://4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132" gracePeriod=30 Nov 23 20:26:17 crc kubenswrapper[4726]: I1123 20:26:17.454121 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 20:26:17 crc kubenswrapper[4726]: I1123 20:26:17.494087 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.216124848 podStartE2EDuration="12.494066546s" podCreationTimestamp="2025-11-23 20:26:05 +0000 UTC" firstStartedPulling="2025-11-23 20:26:12.147201093 +0000 UTC m=+1080.296242049" lastFinishedPulling="2025-11-23 20:26:16.425142791 +0000 UTC m=+1084.574183747" observedRunningTime="2025-11-23 20:26:17.489313069 +0000 UTC m=+1085.638354035" watchObservedRunningTime="2025-11-23 20:26:17.494066546 +0000 UTC m=+1085.643107502" Nov 23 20:26:18 crc kubenswrapper[4726]: I1123 20:26:18.465942 4726 generic.go:334] "Generic (PLEG): container finished" podID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerID="8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd" exitCode=0 Nov 23 20:26:18 crc kubenswrapper[4726]: I1123 20:26:18.466419 4726 generic.go:334] "Generic (PLEG): container finished" podID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerID="1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15" exitCode=2 Nov 23 20:26:18 crc kubenswrapper[4726]: I1123 20:26:18.466434 4726 generic.go:334] "Generic (PLEG): container finished" podID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerID="4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132" exitCode=0 Nov 23 20:26:18 crc kubenswrapper[4726]: I1123 20:26:18.466011 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerDied","Data":"8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd"} Nov 23 20:26:18 crc kubenswrapper[4726]: I1123 20:26:18.466483 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerDied","Data":"1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15"} Nov 23 20:26:18 crc kubenswrapper[4726]: I1123 20:26:18.466504 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerDied","Data":"4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132"} Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.449383 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.525644 4726 generic.go:334] "Generic (PLEG): container finished" podID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerID="83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a" exitCode=0 Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.525838 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.525883 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerDied","Data":"83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a"} Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.528671 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ff1241-f8a5-4595-9ad5-2386fb6a7075","Type":"ContainerDied","Data":"d579eff835a65f2b6774995acb858dc20aa42cb88be9b7fe0f1b0a9bbc586be1"} Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.529066 4726 scope.go:117] "RemoveContainer" containerID="8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.554483 4726 scope.go:117] "RemoveContainer" containerID="1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.574855 4726 scope.go:117] "RemoveContainer" containerID="4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.615132 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-combined-ca-bundle\") pod \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.615220 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-run-httpd\") pod \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.615274 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-config-data\") pod \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.615317 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j26ht\" (UniqueName: \"kubernetes.io/projected/85ff1241-f8a5-4595-9ad5-2386fb6a7075-kube-api-access-j26ht\") pod \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.615431 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-scripts\") pod \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.615493 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-sg-core-conf-yaml\") pod \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.615551 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-log-httpd\") pod \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\" (UID: \"85ff1241-f8a5-4595-9ad5-2386fb6a7075\") " Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.619047 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "85ff1241-f8a5-4595-9ad5-2386fb6a7075" (UID: "85ff1241-f8a5-4595-9ad5-2386fb6a7075"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.619130 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "85ff1241-f8a5-4595-9ad5-2386fb6a7075" (UID: "85ff1241-f8a5-4595-9ad5-2386fb6a7075"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.625412 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-scripts" (OuterVolumeSpecName: "scripts") pod "85ff1241-f8a5-4595-9ad5-2386fb6a7075" (UID: "85ff1241-f8a5-4595-9ad5-2386fb6a7075"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.627146 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85ff1241-f8a5-4595-9ad5-2386fb6a7075-kube-api-access-j26ht" (OuterVolumeSpecName: "kube-api-access-j26ht") pod "85ff1241-f8a5-4595-9ad5-2386fb6a7075" (UID: "85ff1241-f8a5-4595-9ad5-2386fb6a7075"). InnerVolumeSpecName "kube-api-access-j26ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.630470 4726 scope.go:117] "RemoveContainer" containerID="83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.717903 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j26ht\" (UniqueName: \"kubernetes.io/projected/85ff1241-f8a5-4595-9ad5-2386fb6a7075-kube-api-access-j26ht\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.717938 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.717950 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.717961 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ff1241-f8a5-4595-9ad5-2386fb6a7075-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.720617 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85ff1241-f8a5-4595-9ad5-2386fb6a7075" (UID: "85ff1241-f8a5-4595-9ad5-2386fb6a7075"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.721832 4726 scope.go:117] "RemoveContainer" containerID="8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.722483 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd\": container with ID starting with 8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd not found: ID does not exist" containerID="8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.722542 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd"} err="failed to get container status \"8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd\": rpc error: code = NotFound desc = could not find container \"8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd\": container with ID starting with 8efaabc25480f314a0bf90fdc371e9557f97b8490288ce140fd6136190cb22dd not found: ID does not exist" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.722571 4726 scope.go:117] "RemoveContainer" containerID="1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.723070 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15\": container with ID starting with 1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15 not found: ID does not exist" containerID="1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.723110 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15"} err="failed to get container status \"1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15\": rpc error: code = NotFound desc = could not find container \"1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15\": container with ID starting with 1d89c5725d87442ab31458c9c13f653d4e0bc878fe4af2908e0499386b33df15 not found: ID does not exist" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.723138 4726 scope.go:117] "RemoveContainer" containerID="4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.723079 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "85ff1241-f8a5-4595-9ad5-2386fb6a7075" (UID: "85ff1241-f8a5-4595-9ad5-2386fb6a7075"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.723600 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132\": container with ID starting with 4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132 not found: ID does not exist" containerID="4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.723641 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132"} err="failed to get container status \"4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132\": rpc error: code = NotFound desc = could not find container \"4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132\": container with ID starting with 4c16837ccefd64a9b2b73a5c1209b0da603ea6eb5ba4ee64d189daac83ebd132 not found: ID does not exist" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.723664 4726 scope.go:117] "RemoveContainer" containerID="83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.724084 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a\": container with ID starting with 83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a not found: ID does not exist" containerID="83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.724114 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a"} err="failed to get container status \"83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a\": rpc error: code = NotFound desc = could not find container \"83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a\": container with ID starting with 83db638eee989e57a65afdea6857e75c9f36d1cfa980f686d33dfb1a5867615a not found: ID does not exist" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.742105 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-config-data" (OuterVolumeSpecName: "config-data") pod "85ff1241-f8a5-4595-9ad5-2386fb6a7075" (UID: "85ff1241-f8a5-4595-9ad5-2386fb6a7075"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.819253 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.819524 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.819591 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ff1241-f8a5-4595-9ad5-2386fb6a7075-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.862129 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.874879 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.901461 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.901877 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="sg-core" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.901895 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="sg-core" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.901917 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34abb904-d204-4941-afc7-21887ef6703c" containerName="neutron-httpd" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.901924 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="34abb904-d204-4941-afc7-21887ef6703c" containerName="neutron-httpd" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.901943 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="ceilometer-central-agent" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.901950 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="ceilometer-central-agent" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.901967 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="ceilometer-notification-agent" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.901974 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="ceilometer-notification-agent" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.901988 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34abb904-d204-4941-afc7-21887ef6703c" containerName="neutron-api" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.901995 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="34abb904-d204-4941-afc7-21887ef6703c" containerName="neutron-api" Nov 23 20:26:21 crc kubenswrapper[4726]: E1123 20:26:21.902013 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="proxy-httpd" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.902021 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="proxy-httpd" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.902217 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="34abb904-d204-4941-afc7-21887ef6703c" containerName="neutron-httpd" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.902232 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="ceilometer-notification-agent" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.902246 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="ceilometer-central-agent" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.902255 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="sg-core" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.902267 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="34abb904-d204-4941-afc7-21887ef6703c" containerName="neutron-api" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.902287 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" containerName="proxy-httpd" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.904171 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.910218 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.911992 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 20:26:21 crc kubenswrapper[4726]: I1123 20:26:21.924417 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.025623 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-config-data\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.025683 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-log-httpd\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.025918 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.026019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-run-httpd\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.026084 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfk8l\" (UniqueName: \"kubernetes.io/projected/94debbed-92cd-4a0c-bc7a-f85692e013d0-kube-api-access-xfk8l\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.026128 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-scripts\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.026163 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.127994 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-config-data\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.128061 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-log-httpd\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.128106 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.128131 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-run-httpd\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.128158 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfk8l\" (UniqueName: \"kubernetes.io/projected/94debbed-92cd-4a0c-bc7a-f85692e013d0-kube-api-access-xfk8l\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.128181 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.128200 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-scripts\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.128757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-run-httpd\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.129201 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-log-httpd\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.132956 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.134377 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-config-data\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.136506 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-scripts\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.148559 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.149709 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfk8l\" (UniqueName: \"kubernetes.io/projected/94debbed-92cd-4a0c-bc7a-f85692e013d0-kube-api-access-xfk8l\") pod \"ceilometer-0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.198393 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-45xfq"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.199578 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.211044 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-45xfq"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.223458 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.317290 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-bw8f5"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.318689 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.338528 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g2rf\" (UniqueName: \"kubernetes.io/projected/bb3fda45-1598-4496-9463-89d4f1cc2ef5-kube-api-access-6g2rf\") pod \"nova-api-db-create-45xfq\" (UID: \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\") " pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.338772 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb3fda45-1598-4496-9463-89d4f1cc2ef5-operator-scripts\") pod \"nova-api-db-create-45xfq\" (UID: \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\") " pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.341490 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bw8f5"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.350104 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8822-account-create-fxsgq"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.351425 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.358297 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8822-account-create-fxsgq"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.361771 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.440977 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w86sc\" (UniqueName: \"kubernetes.io/projected/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-kube-api-access-w86sc\") pod \"nova-cell0-db-create-bw8f5\" (UID: \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\") " pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.442590 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-operator-scripts\") pod \"nova-cell0-db-create-bw8f5\" (UID: \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\") " pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.442693 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g2rf\" (UniqueName: \"kubernetes.io/projected/bb3fda45-1598-4496-9463-89d4f1cc2ef5-kube-api-access-6g2rf\") pod \"nova-api-db-create-45xfq\" (UID: \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\") " pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.442731 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb3fda45-1598-4496-9463-89d4f1cc2ef5-operator-scripts\") pod \"nova-api-db-create-45xfq\" (UID: \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\") " pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.445312 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb3fda45-1598-4496-9463-89d4f1cc2ef5-operator-scripts\") pod \"nova-api-db-create-45xfq\" (UID: \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\") " pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.467691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g2rf\" (UniqueName: \"kubernetes.io/projected/bb3fda45-1598-4496-9463-89d4f1cc2ef5-kube-api-access-6g2rf\") pod \"nova-api-db-create-45xfq\" (UID: \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\") " pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.512571 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xbng7"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.513954 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.524810 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-06fb-account-create-s7l7m"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.526687 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.540743 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.544317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w86sc\" (UniqueName: \"kubernetes.io/projected/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-kube-api-access-w86sc\") pod \"nova-cell0-db-create-bw8f5\" (UID: \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\") " pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.544350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-operator-scripts\") pod \"nova-cell0-db-create-bw8f5\" (UID: \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\") " pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.544391 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kbmh\" (UniqueName: \"kubernetes.io/projected/79707a2d-f705-455f-a097-50c2e30eed2f-kube-api-access-4kbmh\") pod \"nova-api-8822-account-create-fxsgq\" (UID: \"79707a2d-f705-455f-a097-50c2e30eed2f\") " pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.544443 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79707a2d-f705-455f-a097-50c2e30eed2f-operator-scripts\") pod \"nova-api-8822-account-create-fxsgq\" (UID: \"79707a2d-f705-455f-a097-50c2e30eed2f\") " pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.545758 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-operator-scripts\") pod \"nova-cell0-db-create-bw8f5\" (UID: \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\") " pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.546059 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xbng7"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.548040 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.557371 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-06fb-account-create-s7l7m"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.570658 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w86sc\" (UniqueName: \"kubernetes.io/projected/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-kube-api-access-w86sc\") pod \"nova-cell0-db-create-bw8f5\" (UID: \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\") " pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.603780 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85ff1241-f8a5-4595-9ad5-2386fb6a7075" path="/var/lib/kubelet/pods/85ff1241-f8a5-4595-9ad5-2386fb6a7075/volumes" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.648287 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kbmh\" (UniqueName: \"kubernetes.io/projected/79707a2d-f705-455f-a097-50c2e30eed2f-kube-api-access-4kbmh\") pod \"nova-api-8822-account-create-fxsgq\" (UID: \"79707a2d-f705-455f-a097-50c2e30eed2f\") " pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.649167 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rqjj\" (UniqueName: \"kubernetes.io/projected/08af9ecf-3570-4e1c-98bd-163b5116634a-kube-api-access-4rqjj\") pod \"nova-cell0-06fb-account-create-s7l7m\" (UID: \"08af9ecf-3570-4e1c-98bd-163b5116634a\") " pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.649327 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk8v4\" (UniqueName: \"kubernetes.io/projected/3707a249-9a42-4ee2-8f07-c4a72c295a46-kube-api-access-tk8v4\") pod \"nova-cell1-db-create-xbng7\" (UID: \"3707a249-9a42-4ee2-8f07-c4a72c295a46\") " pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.649462 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79707a2d-f705-455f-a097-50c2e30eed2f-operator-scripts\") pod \"nova-api-8822-account-create-fxsgq\" (UID: \"79707a2d-f705-455f-a097-50c2e30eed2f\") " pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.650457 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08af9ecf-3570-4e1c-98bd-163b5116634a-operator-scripts\") pod \"nova-cell0-06fb-account-create-s7l7m\" (UID: \"08af9ecf-3570-4e1c-98bd-163b5116634a\") " pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.650571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3707a249-9a42-4ee2-8f07-c4a72c295a46-operator-scripts\") pod \"nova-cell1-db-create-xbng7\" (UID: \"3707a249-9a42-4ee2-8f07-c4a72c295a46\") " pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.650627 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.650390 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79707a2d-f705-455f-a097-50c2e30eed2f-operator-scripts\") pod \"nova-api-8822-account-create-fxsgq\" (UID: \"79707a2d-f705-455f-a097-50c2e30eed2f\") " pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.674541 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kbmh\" (UniqueName: \"kubernetes.io/projected/79707a2d-f705-455f-a097-50c2e30eed2f-kube-api-access-4kbmh\") pod \"nova-api-8822-account-create-fxsgq\" (UID: \"79707a2d-f705-455f-a097-50c2e30eed2f\") " pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.717133 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-186c-account-create-x2lmc"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.729383 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.736363 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.747280 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-186c-account-create-x2lmc"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.758427 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08af9ecf-3570-4e1c-98bd-163b5116634a-operator-scripts\") pod \"nova-cell0-06fb-account-create-s7l7m\" (UID: \"08af9ecf-3570-4e1c-98bd-163b5116634a\") " pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.758484 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3707a249-9a42-4ee2-8f07-c4a72c295a46-operator-scripts\") pod \"nova-cell1-db-create-xbng7\" (UID: \"3707a249-9a42-4ee2-8f07-c4a72c295a46\") " pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.758578 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rqjj\" (UniqueName: \"kubernetes.io/projected/08af9ecf-3570-4e1c-98bd-163b5116634a-kube-api-access-4rqjj\") pod \"nova-cell0-06fb-account-create-s7l7m\" (UID: \"08af9ecf-3570-4e1c-98bd-163b5116634a\") " pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.758611 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk8v4\" (UniqueName: \"kubernetes.io/projected/3707a249-9a42-4ee2-8f07-c4a72c295a46-kube-api-access-tk8v4\") pod \"nova-cell1-db-create-xbng7\" (UID: \"3707a249-9a42-4ee2-8f07-c4a72c295a46\") " pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.759558 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08af9ecf-3570-4e1c-98bd-163b5116634a-operator-scripts\") pod \"nova-cell0-06fb-account-create-s7l7m\" (UID: \"08af9ecf-3570-4e1c-98bd-163b5116634a\") " pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.760182 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3707a249-9a42-4ee2-8f07-c4a72c295a46-operator-scripts\") pod \"nova-cell1-db-create-xbng7\" (UID: \"3707a249-9a42-4ee2-8f07-c4a72c295a46\") " pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.787007 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rqjj\" (UniqueName: \"kubernetes.io/projected/08af9ecf-3570-4e1c-98bd-163b5116634a-kube-api-access-4rqjj\") pod \"nova-cell0-06fb-account-create-s7l7m\" (UID: \"08af9ecf-3570-4e1c-98bd-163b5116634a\") " pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.787795 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk8v4\" (UniqueName: \"kubernetes.io/projected/3707a249-9a42-4ee2-8f07-c4a72c295a46-kube-api-access-tk8v4\") pod \"nova-cell1-db-create-xbng7\" (UID: \"3707a249-9a42-4ee2-8f07-c4a72c295a46\") " pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.853125 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.861021 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100d7602-9a19-4155-b38b-f7bfd4b685ad-operator-scripts\") pod \"nova-cell1-186c-account-create-x2lmc\" (UID: \"100d7602-9a19-4155-b38b-f7bfd4b685ad\") " pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.863090 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxp4l\" (UniqueName: \"kubernetes.io/projected/100d7602-9a19-4155-b38b-f7bfd4b685ad-kube-api-access-bxp4l\") pod \"nova-cell1-186c-account-create-x2lmc\" (UID: \"100d7602-9a19-4155-b38b-f7bfd4b685ad\") " pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.868280 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.874601 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.965043 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100d7602-9a19-4155-b38b-f7bfd4b685ad-operator-scripts\") pod \"nova-cell1-186c-account-create-x2lmc\" (UID: \"100d7602-9a19-4155-b38b-f7bfd4b685ad\") " pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.965171 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxp4l\" (UniqueName: \"kubernetes.io/projected/100d7602-9a19-4155-b38b-f7bfd4b685ad-kube-api-access-bxp4l\") pod \"nova-cell1-186c-account-create-x2lmc\" (UID: \"100d7602-9a19-4155-b38b-f7bfd4b685ad\") " pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.966277 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100d7602-9a19-4155-b38b-f7bfd4b685ad-operator-scripts\") pod \"nova-cell1-186c-account-create-x2lmc\" (UID: \"100d7602-9a19-4155-b38b-f7bfd4b685ad\") " pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.972432 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:22 crc kubenswrapper[4726]: I1123 20:26:22.982539 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxp4l\" (UniqueName: \"kubernetes.io/projected/100d7602-9a19-4155-b38b-f7bfd4b685ad-kube-api-access-bxp4l\") pod \"nova-cell1-186c-account-create-x2lmc\" (UID: \"100d7602-9a19-4155-b38b-f7bfd4b685ad\") " pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.056407 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.222411 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-45xfq"] Nov 23 20:26:23 crc kubenswrapper[4726]: W1123 20:26:23.269566 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb3fda45_1598_4496_9463_89d4f1cc2ef5.slice/crio-adcf4530950d94710566139cf7637267a6fc4017cea61ed736b263b2bd248021 WatchSource:0}: Error finding container adcf4530950d94710566139cf7637267a6fc4017cea61ed736b263b2bd248021: Status 404 returned error can't find the container with id adcf4530950d94710566139cf7637267a6fc4017cea61ed736b263b2bd248021 Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.301025 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.364510 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bw8f5"] Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.537699 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xbng7"] Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.629198 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bw8f5" event={"ID":"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2","Type":"ContainerStarted","Data":"4a7c16657d60498d323e5c104b0777cebd653eaa6ac184618c2eca228fc6661c"} Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.661342 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerStarted","Data":"c880915603f8a8a9f2d4d2a9360f4e473053c284c7f11059ce627879ea63e625"} Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.666775 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-45xfq" event={"ID":"bb3fda45-1598-4496-9463-89d4f1cc2ef5","Type":"ContainerStarted","Data":"adcf4530950d94710566139cf7637267a6fc4017cea61ed736b263b2bd248021"} Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.728546 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-06fb-account-create-s7l7m"] Nov 23 20:26:23 crc kubenswrapper[4726]: I1123 20:26:23.742194 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8822-account-create-fxsgq"] Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.001411 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-186c-account-create-x2lmc"] Nov 23 20:26:24 crc kubenswrapper[4726]: W1123 20:26:24.026043 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod100d7602_9a19_4155_b38b_f7bfd4b685ad.slice/crio-8bb1bd2aec87ccac9a764352629ac7ce1667e259b45cef50c3edc40066ec5675 WatchSource:0}: Error finding container 8bb1bd2aec87ccac9a764352629ac7ce1667e259b45cef50c3edc40066ec5675: Status 404 returned error can't find the container with id 8bb1bd2aec87ccac9a764352629ac7ce1667e259b45cef50c3edc40066ec5675 Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.678078 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerStarted","Data":"9df51064fc9d4ce90560cb54e4dcfaf2d9ec35c60f07f233fae2fc11cd9b013c"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.686036 4726 generic.go:334] "Generic (PLEG): container finished" podID="100d7602-9a19-4155-b38b-f7bfd4b685ad" containerID="0774a1b0a35dd1676ffe601d3b108f18550694efa16a719335cd55dfca76d765" exitCode=0 Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.686233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-186c-account-create-x2lmc" event={"ID":"100d7602-9a19-4155-b38b-f7bfd4b685ad","Type":"ContainerDied","Data":"0774a1b0a35dd1676ffe601d3b108f18550694efa16a719335cd55dfca76d765"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.686741 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-186c-account-create-x2lmc" event={"ID":"100d7602-9a19-4155-b38b-f7bfd4b685ad","Type":"ContainerStarted","Data":"8bb1bd2aec87ccac9a764352629ac7ce1667e259b45cef50c3edc40066ec5675"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.688309 4726 generic.go:334] "Generic (PLEG): container finished" podID="bb3fda45-1598-4496-9463-89d4f1cc2ef5" containerID="1dab6608897606fd5c259c5c7de8625df3c33a89e7745f46c17a4b89fed137d6" exitCode=0 Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.688417 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-45xfq" event={"ID":"bb3fda45-1598-4496-9463-89d4f1cc2ef5","Type":"ContainerDied","Data":"1dab6608897606fd5c259c5c7de8625df3c33a89e7745f46c17a4b89fed137d6"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.690411 4726 generic.go:334] "Generic (PLEG): container finished" podID="3707a249-9a42-4ee2-8f07-c4a72c295a46" containerID="cd4da0aef4c26bd531f7b1d20e1f587bf7ba4109c52ea55ae21d0296fc521d6f" exitCode=0 Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.690511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xbng7" event={"ID":"3707a249-9a42-4ee2-8f07-c4a72c295a46","Type":"ContainerDied","Data":"cd4da0aef4c26bd531f7b1d20e1f587bf7ba4109c52ea55ae21d0296fc521d6f"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.690590 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xbng7" event={"ID":"3707a249-9a42-4ee2-8f07-c4a72c295a46","Type":"ContainerStarted","Data":"4731350f8c17627990a4aa14cf6a802545e631d8f5c5e8f2cb31e8ebae6ac6f2"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.691768 4726 generic.go:334] "Generic (PLEG): container finished" podID="a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2" containerID="ceebb89ddb0e75b4260a3d34ccc2eedc24c0febca6e177d82bda7d26ac1722a8" exitCode=0 Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.691983 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bw8f5" event={"ID":"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2","Type":"ContainerDied","Data":"ceebb89ddb0e75b4260a3d34ccc2eedc24c0febca6e177d82bda7d26ac1722a8"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.695196 4726 generic.go:334] "Generic (PLEG): container finished" podID="08af9ecf-3570-4e1c-98bd-163b5116634a" containerID="05f0b309c0be90990868e64f86abfdda00d40fe6b71cd616e2151dd989f0bde5" exitCode=0 Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.695341 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-06fb-account-create-s7l7m" event={"ID":"08af9ecf-3570-4e1c-98bd-163b5116634a","Type":"ContainerDied","Data":"05f0b309c0be90990868e64f86abfdda00d40fe6b71cd616e2151dd989f0bde5"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.695431 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-06fb-account-create-s7l7m" event={"ID":"08af9ecf-3570-4e1c-98bd-163b5116634a","Type":"ContainerStarted","Data":"914415ae6179938ffb945a4e0388773257edca609d6a75e456937aee316ffc93"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.697187 4726 generic.go:334] "Generic (PLEG): container finished" podID="79707a2d-f705-455f-a097-50c2e30eed2f" containerID="aa5df2feac6a8cc66913a2834fbc6eb25d7d8a29c868d7b1de754e0a20640773" exitCode=0 Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.697318 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8822-account-create-fxsgq" event={"ID":"79707a2d-f705-455f-a097-50c2e30eed2f","Type":"ContainerDied","Data":"aa5df2feac6a8cc66913a2834fbc6eb25d7d8a29c868d7b1de754e0a20640773"} Nov 23 20:26:24 crc kubenswrapper[4726]: I1123 20:26:24.697415 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8822-account-create-fxsgq" event={"ID":"79707a2d-f705-455f-a097-50c2e30eed2f","Type":"ContainerStarted","Data":"a4bdb6fa77d9f33548b09cac63d84c3547c15e0d54288cf8d496f90f7b7d4a3a"} Nov 23 20:26:25 crc kubenswrapper[4726]: I1123 20:26:25.709719 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerStarted","Data":"c1c44993720e8bfe84abc72ead753ff54c098bc2f4c92bef50223afbc08af850"} Nov 23 20:26:25 crc kubenswrapper[4726]: I1123 20:26:25.710289 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerStarted","Data":"10ab5f69ff760f73f6a759bee2a03a28da55c14b547d6fa0d20f9365c38ceb00"} Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.160812 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.294163 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxp4l\" (UniqueName: \"kubernetes.io/projected/100d7602-9a19-4155-b38b-f7bfd4b685ad-kube-api-access-bxp4l\") pod \"100d7602-9a19-4155-b38b-f7bfd4b685ad\" (UID: \"100d7602-9a19-4155-b38b-f7bfd4b685ad\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.294729 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100d7602-9a19-4155-b38b-f7bfd4b685ad-operator-scripts\") pod \"100d7602-9a19-4155-b38b-f7bfd4b685ad\" (UID: \"100d7602-9a19-4155-b38b-f7bfd4b685ad\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.363139 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/100d7602-9a19-4155-b38b-f7bfd4b685ad-kube-api-access-bxp4l" (OuterVolumeSpecName: "kube-api-access-bxp4l") pod "100d7602-9a19-4155-b38b-f7bfd4b685ad" (UID: "100d7602-9a19-4155-b38b-f7bfd4b685ad"). InnerVolumeSpecName "kube-api-access-bxp4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.376531 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/100d7602-9a19-4155-b38b-f7bfd4b685ad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "100d7602-9a19-4155-b38b-f7bfd4b685ad" (UID: "100d7602-9a19-4155-b38b-f7bfd4b685ad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.401356 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxp4l\" (UniqueName: \"kubernetes.io/projected/100d7602-9a19-4155-b38b-f7bfd4b685ad-kube-api-access-bxp4l\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.401394 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100d7602-9a19-4155-b38b-f7bfd4b685ad-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.626059 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.630995 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.641357 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.651482 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.669191 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.711368 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79707a2d-f705-455f-a097-50c2e30eed2f-operator-scripts\") pod \"79707a2d-f705-455f-a097-50c2e30eed2f\" (UID: \"79707a2d-f705-455f-a097-50c2e30eed2f\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.711437 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3707a249-9a42-4ee2-8f07-c4a72c295a46-operator-scripts\") pod \"3707a249-9a42-4ee2-8f07-c4a72c295a46\" (UID: \"3707a249-9a42-4ee2-8f07-c4a72c295a46\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.711609 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rqjj\" (UniqueName: \"kubernetes.io/projected/08af9ecf-3570-4e1c-98bd-163b5116634a-kube-api-access-4rqjj\") pod \"08af9ecf-3570-4e1c-98bd-163b5116634a\" (UID: \"08af9ecf-3570-4e1c-98bd-163b5116634a\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.711665 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk8v4\" (UniqueName: \"kubernetes.io/projected/3707a249-9a42-4ee2-8f07-c4a72c295a46-kube-api-access-tk8v4\") pod \"3707a249-9a42-4ee2-8f07-c4a72c295a46\" (UID: \"3707a249-9a42-4ee2-8f07-c4a72c295a46\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.711689 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kbmh\" (UniqueName: \"kubernetes.io/projected/79707a2d-f705-455f-a097-50c2e30eed2f-kube-api-access-4kbmh\") pod \"79707a2d-f705-455f-a097-50c2e30eed2f\" (UID: \"79707a2d-f705-455f-a097-50c2e30eed2f\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.711729 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08af9ecf-3570-4e1c-98bd-163b5116634a-operator-scripts\") pod \"08af9ecf-3570-4e1c-98bd-163b5116634a\" (UID: \"08af9ecf-3570-4e1c-98bd-163b5116634a\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.712474 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79707a2d-f705-455f-a097-50c2e30eed2f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79707a2d-f705-455f-a097-50c2e30eed2f" (UID: "79707a2d-f705-455f-a097-50c2e30eed2f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.712479 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3707a249-9a42-4ee2-8f07-c4a72c295a46-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3707a249-9a42-4ee2-8f07-c4a72c295a46" (UID: "3707a249-9a42-4ee2-8f07-c4a72c295a46"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.713815 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08af9ecf-3570-4e1c-98bd-163b5116634a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "08af9ecf-3570-4e1c-98bd-163b5116634a" (UID: "08af9ecf-3570-4e1c-98bd-163b5116634a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.719661 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3707a249-9a42-4ee2-8f07-c4a72c295a46-kube-api-access-tk8v4" (OuterVolumeSpecName: "kube-api-access-tk8v4") pod "3707a249-9a42-4ee2-8f07-c4a72c295a46" (UID: "3707a249-9a42-4ee2-8f07-c4a72c295a46"). InnerVolumeSpecName "kube-api-access-tk8v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.721034 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79707a2d-f705-455f-a097-50c2e30eed2f-kube-api-access-4kbmh" (OuterVolumeSpecName: "kube-api-access-4kbmh") pod "79707a2d-f705-455f-a097-50c2e30eed2f" (UID: "79707a2d-f705-455f-a097-50c2e30eed2f"). InnerVolumeSpecName "kube-api-access-4kbmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.722365 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08af9ecf-3570-4e1c-98bd-163b5116634a-kube-api-access-4rqjj" (OuterVolumeSpecName: "kube-api-access-4rqjj") pod "08af9ecf-3570-4e1c-98bd-163b5116634a" (UID: "08af9ecf-3570-4e1c-98bd-163b5116634a"). InnerVolumeSpecName "kube-api-access-4rqjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.730985 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-06fb-account-create-s7l7m" event={"ID":"08af9ecf-3570-4e1c-98bd-163b5116634a","Type":"ContainerDied","Data":"914415ae6179938ffb945a4e0388773257edca609d6a75e456937aee316ffc93"} Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.731034 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="914415ae6179938ffb945a4e0388773257edca609d6a75e456937aee316ffc93" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.731121 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-06fb-account-create-s7l7m" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.742554 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8822-account-create-fxsgq" event={"ID":"79707a2d-f705-455f-a097-50c2e30eed2f","Type":"ContainerDied","Data":"a4bdb6fa77d9f33548b09cac63d84c3547c15e0d54288cf8d496f90f7b7d4a3a"} Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.742610 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4bdb6fa77d9f33548b09cac63d84c3547c15e0d54288cf8d496f90f7b7d4a3a" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.742675 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8822-account-create-fxsgq" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.750402 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-186c-account-create-x2lmc" event={"ID":"100d7602-9a19-4155-b38b-f7bfd4b685ad","Type":"ContainerDied","Data":"8bb1bd2aec87ccac9a764352629ac7ce1667e259b45cef50c3edc40066ec5675"} Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.750449 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bb1bd2aec87ccac9a764352629ac7ce1667e259b45cef50c3edc40066ec5675" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.750564 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-186c-account-create-x2lmc" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.760643 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-45xfq" event={"ID":"bb3fda45-1598-4496-9463-89d4f1cc2ef5","Type":"ContainerDied","Data":"adcf4530950d94710566139cf7637267a6fc4017cea61ed736b263b2bd248021"} Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.760685 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adcf4530950d94710566139cf7637267a6fc4017cea61ed736b263b2bd248021" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.760779 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-45xfq" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.765712 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xbng7" event={"ID":"3707a249-9a42-4ee2-8f07-c4a72c295a46","Type":"ContainerDied","Data":"4731350f8c17627990a4aa14cf6a802545e631d8f5c5e8f2cb31e8ebae6ac6f2"} Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.765754 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4731350f8c17627990a4aa14cf6a802545e631d8f5c5e8f2cb31e8ebae6ac6f2" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.765836 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xbng7" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.780387 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bw8f5" event={"ID":"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2","Type":"ContainerDied","Data":"4a7c16657d60498d323e5c104b0777cebd653eaa6ac184618c2eca228fc6661c"} Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.780429 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a7c16657d60498d323e5c104b0777cebd653eaa6ac184618c2eca228fc6661c" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.780491 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bw8f5" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.814031 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-operator-scripts\") pod \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\" (UID: \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.814074 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w86sc\" (UniqueName: \"kubernetes.io/projected/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-kube-api-access-w86sc\") pod \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\" (UID: \"a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.814202 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g2rf\" (UniqueName: \"kubernetes.io/projected/bb3fda45-1598-4496-9463-89d4f1cc2ef5-kube-api-access-6g2rf\") pod \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\" (UID: \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815231 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb3fda45-1598-4496-9463-89d4f1cc2ef5-operator-scripts\") pod \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\" (UID: \"bb3fda45-1598-4496-9463-89d4f1cc2ef5\") " Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815346 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2" (UID: "a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815643 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rqjj\" (UniqueName: \"kubernetes.io/projected/08af9ecf-3570-4e1c-98bd-163b5116634a-kube-api-access-4rqjj\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815659 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk8v4\" (UniqueName: \"kubernetes.io/projected/3707a249-9a42-4ee2-8f07-c4a72c295a46-kube-api-access-tk8v4\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815670 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kbmh\" (UniqueName: \"kubernetes.io/projected/79707a2d-f705-455f-a097-50c2e30eed2f-kube-api-access-4kbmh\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815680 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08af9ecf-3570-4e1c-98bd-163b5116634a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815689 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815698 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79707a2d-f705-455f-a097-50c2e30eed2f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815708 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3707a249-9a42-4ee2-8f07-c4a72c295a46-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.815804 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb3fda45-1598-4496-9463-89d4f1cc2ef5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bb3fda45-1598-4496-9463-89d4f1cc2ef5" (UID: "bb3fda45-1598-4496-9463-89d4f1cc2ef5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.818055 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-kube-api-access-w86sc" (OuterVolumeSpecName: "kube-api-access-w86sc") pod "a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2" (UID: "a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2"). InnerVolumeSpecName "kube-api-access-w86sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.821000 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb3fda45-1598-4496-9463-89d4f1cc2ef5-kube-api-access-6g2rf" (OuterVolumeSpecName: "kube-api-access-6g2rf") pod "bb3fda45-1598-4496-9463-89d4f1cc2ef5" (UID: "bb3fda45-1598-4496-9463-89d4f1cc2ef5"). InnerVolumeSpecName "kube-api-access-6g2rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.917542 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w86sc\" (UniqueName: \"kubernetes.io/projected/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2-kube-api-access-w86sc\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.917578 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g2rf\" (UniqueName: \"kubernetes.io/projected/bb3fda45-1598-4496-9463-89d4f1cc2ef5-kube-api-access-6g2rf\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:26 crc kubenswrapper[4726]: I1123 20:26:26.917592 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb3fda45-1598-4496-9463-89d4f1cc2ef5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:27 crc kubenswrapper[4726]: I1123 20:26:27.792739 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerStarted","Data":"1d19b5df6e99c0abcd89710424fc852a724937af1d0b2047b2044e4c7a25f1ce"} Nov 23 20:26:27 crc kubenswrapper[4726]: I1123 20:26:27.793392 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 20:26:27 crc kubenswrapper[4726]: I1123 20:26:27.793187 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="proxy-httpd" containerID="cri-o://1d19b5df6e99c0abcd89710424fc852a724937af1d0b2047b2044e4c7a25f1ce" gracePeriod=30 Nov 23 20:26:27 crc kubenswrapper[4726]: I1123 20:26:27.793204 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="sg-core" containerID="cri-o://c1c44993720e8bfe84abc72ead753ff54c098bc2f4c92bef50223afbc08af850" gracePeriod=30 Nov 23 20:26:27 crc kubenswrapper[4726]: I1123 20:26:27.793217 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="ceilometer-notification-agent" containerID="cri-o://10ab5f69ff760f73f6a759bee2a03a28da55c14b547d6fa0d20f9365c38ceb00" gracePeriod=30 Nov 23 20:26:27 crc kubenswrapper[4726]: I1123 20:26:27.792941 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="ceilometer-central-agent" containerID="cri-o://9df51064fc9d4ce90560cb54e4dcfaf2d9ec35c60f07f233fae2fc11cd9b013c" gracePeriod=30 Nov 23 20:26:27 crc kubenswrapper[4726]: I1123 20:26:27.818759 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.80391371 podStartE2EDuration="6.818732726s" podCreationTimestamp="2025-11-23 20:26:21 +0000 UTC" firstStartedPulling="2025-11-23 20:26:22.890072567 +0000 UTC m=+1091.039113523" lastFinishedPulling="2025-11-23 20:26:26.904891583 +0000 UTC m=+1095.053932539" observedRunningTime="2025-11-23 20:26:27.816674511 +0000 UTC m=+1095.965715467" watchObservedRunningTime="2025-11-23 20:26:27.818732726 +0000 UTC m=+1095.967773682" Nov 23 20:26:28 crc kubenswrapper[4726]: I1123 20:26:28.840652 4726 generic.go:334] "Generic (PLEG): container finished" podID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerID="1d19b5df6e99c0abcd89710424fc852a724937af1d0b2047b2044e4c7a25f1ce" exitCode=0 Nov 23 20:26:28 crc kubenswrapper[4726]: I1123 20:26:28.841145 4726 generic.go:334] "Generic (PLEG): container finished" podID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerID="c1c44993720e8bfe84abc72ead753ff54c098bc2f4c92bef50223afbc08af850" exitCode=2 Nov 23 20:26:28 crc kubenswrapper[4726]: I1123 20:26:28.841156 4726 generic.go:334] "Generic (PLEG): container finished" podID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerID="10ab5f69ff760f73f6a759bee2a03a28da55c14b547d6fa0d20f9365c38ceb00" exitCode=0 Nov 23 20:26:28 crc kubenswrapper[4726]: I1123 20:26:28.840926 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerDied","Data":"1d19b5df6e99c0abcd89710424fc852a724937af1d0b2047b2044e4c7a25f1ce"} Nov 23 20:26:28 crc kubenswrapper[4726]: I1123 20:26:28.841215 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerDied","Data":"c1c44993720e8bfe84abc72ead753ff54c098bc2f4c92bef50223afbc08af850"} Nov 23 20:26:28 crc kubenswrapper[4726]: I1123 20:26:28.841232 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerDied","Data":"10ab5f69ff760f73f6a759bee2a03a28da55c14b547d6fa0d20f9365c38ceb00"} Nov 23 20:26:31 crc kubenswrapper[4726]: I1123 20:26:31.875334 4726 generic.go:334] "Generic (PLEG): container finished" podID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerID="9df51064fc9d4ce90560cb54e4dcfaf2d9ec35c60f07f233fae2fc11cd9b013c" exitCode=0 Nov 23 20:26:31 crc kubenswrapper[4726]: I1123 20:26:31.876269 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerDied","Data":"9df51064fc9d4ce90560cb54e4dcfaf2d9ec35c60f07f233fae2fc11cd9b013c"} Nov 23 20:26:31 crc kubenswrapper[4726]: I1123 20:26:31.876324 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94debbed-92cd-4a0c-bc7a-f85692e013d0","Type":"ContainerDied","Data":"c880915603f8a8a9f2d4d2a9360f4e473053c284c7f11059ce627879ea63e625"} Nov 23 20:26:31 crc kubenswrapper[4726]: I1123 20:26:31.876341 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c880915603f8a8a9f2d4d2a9360f4e473053c284c7f11059ce627879ea63e625" Nov 23 20:26:31 crc kubenswrapper[4726]: I1123 20:26:31.911402 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.048737 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-config-data\") pod \"94debbed-92cd-4a0c-bc7a-f85692e013d0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.050236 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-combined-ca-bundle\") pod \"94debbed-92cd-4a0c-bc7a-f85692e013d0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.050414 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfk8l\" (UniqueName: \"kubernetes.io/projected/94debbed-92cd-4a0c-bc7a-f85692e013d0-kube-api-access-xfk8l\") pod \"94debbed-92cd-4a0c-bc7a-f85692e013d0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.050516 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-log-httpd\") pod \"94debbed-92cd-4a0c-bc7a-f85692e013d0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.050710 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-run-httpd\") pod \"94debbed-92cd-4a0c-bc7a-f85692e013d0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.050750 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-sg-core-conf-yaml\") pod \"94debbed-92cd-4a0c-bc7a-f85692e013d0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.050774 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-scripts\") pod \"94debbed-92cd-4a0c-bc7a-f85692e013d0\" (UID: \"94debbed-92cd-4a0c-bc7a-f85692e013d0\") " Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.051538 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "94debbed-92cd-4a0c-bc7a-f85692e013d0" (UID: "94debbed-92cd-4a0c-bc7a-f85692e013d0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.051856 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "94debbed-92cd-4a0c-bc7a-f85692e013d0" (UID: "94debbed-92cd-4a0c-bc7a-f85692e013d0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.053009 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.053043 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94debbed-92cd-4a0c-bc7a-f85692e013d0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.070723 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94debbed-92cd-4a0c-bc7a-f85692e013d0-kube-api-access-xfk8l" (OuterVolumeSpecName: "kube-api-access-xfk8l") pod "94debbed-92cd-4a0c-bc7a-f85692e013d0" (UID: "94debbed-92cd-4a0c-bc7a-f85692e013d0"). InnerVolumeSpecName "kube-api-access-xfk8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.071709 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-scripts" (OuterVolumeSpecName: "scripts") pod "94debbed-92cd-4a0c-bc7a-f85692e013d0" (UID: "94debbed-92cd-4a0c-bc7a-f85692e013d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.078365 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "94debbed-92cd-4a0c-bc7a-f85692e013d0" (UID: "94debbed-92cd-4a0c-bc7a-f85692e013d0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.134021 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94debbed-92cd-4a0c-bc7a-f85692e013d0" (UID: "94debbed-92cd-4a0c-bc7a-f85692e013d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.149594 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-config-data" (OuterVolumeSpecName: "config-data") pod "94debbed-92cd-4a0c-bc7a-f85692e013d0" (UID: "94debbed-92cd-4a0c-bc7a-f85692e013d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.155202 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.155239 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.155253 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfk8l\" (UniqueName: \"kubernetes.io/projected/94debbed-92cd-4a0c-bc7a-f85692e013d0-kube-api-access-xfk8l\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.155266 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.155279 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94debbed-92cd-4a0c-bc7a-f85692e013d0-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.885595 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.910653 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.920048 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946026 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946499 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="ceilometer-central-agent" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946524 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="ceilometer-central-agent" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946540 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="100d7602-9a19-4155-b38b-f7bfd4b685ad" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946549 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="100d7602-9a19-4155-b38b-f7bfd4b685ad" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946581 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79707a2d-f705-455f-a097-50c2e30eed2f" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946590 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="79707a2d-f705-455f-a097-50c2e30eed2f" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946606 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3707a249-9a42-4ee2-8f07-c4a72c295a46" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946616 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3707a249-9a42-4ee2-8f07-c4a72c295a46" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946635 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08af9ecf-3570-4e1c-98bd-163b5116634a" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946644 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="08af9ecf-3570-4e1c-98bd-163b5116634a" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946669 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946676 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946687 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="proxy-httpd" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946694 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="proxy-httpd" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946709 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3fda45-1598-4496-9463-89d4f1cc2ef5" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946717 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3fda45-1598-4496-9463-89d4f1cc2ef5" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946728 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="ceilometer-notification-agent" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946735 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="ceilometer-notification-agent" Nov 23 20:26:32 crc kubenswrapper[4726]: E1123 20:26:32.946749 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="sg-core" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946757 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="sg-core" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.946984 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="ceilometer-central-agent" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947005 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb3fda45-1598-4496-9463-89d4f1cc2ef5" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947017 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3707a249-9a42-4ee2-8f07-c4a72c295a46" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947028 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="sg-core" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947040 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="79707a2d-f705-455f-a097-50c2e30eed2f" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947052 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="ceilometer-notification-agent" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947063 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2" containerName="mariadb-database-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947075 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="100d7602-9a19-4155-b38b-f7bfd4b685ad" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947087 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="08af9ecf-3570-4e1c-98bd-163b5116634a" containerName="mariadb-account-create" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.947101 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" containerName="proxy-httpd" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.948826 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.952377 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.952554 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 20:26:32 crc kubenswrapper[4726]: I1123 20:26:32.957466 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.060639 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z4vqv"] Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.061815 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.064376 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-w42x6" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.064725 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.065684 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.072437 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-scripts\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.072599 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.072729 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srf4v\" (UniqueName: \"kubernetes.io/projected/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-kube-api-access-srf4v\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.072834 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-config-data\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.072951 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.073051 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-log-httpd\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.073140 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-run-httpd\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.074124 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z4vqv"] Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175369 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-config-data\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175459 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srf4v\" (UniqueName: \"kubernetes.io/projected/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-kube-api-access-srf4v\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175496 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mk5v\" (UniqueName: \"kubernetes.io/projected/094675fb-839b-47fe-b1cd-dec2fafbf906-kube-api-access-6mk5v\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175528 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-scripts\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175553 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175596 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-config-data\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175625 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175670 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-log-httpd\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175702 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-run-httpd\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175749 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-scripts\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.175780 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.177681 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-run-httpd\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.177690 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-log-httpd\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.182274 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-scripts\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.182758 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.182894 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.183713 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-config-data\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.197691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srf4v\" (UniqueName: \"kubernetes.io/projected/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-kube-api-access-srf4v\") pod \"ceilometer-0\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.267388 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.277232 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-config-data\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.277279 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mk5v\" (UniqueName: \"kubernetes.io/projected/094675fb-839b-47fe-b1cd-dec2fafbf906-kube-api-access-6mk5v\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.277306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-scripts\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.277332 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.280980 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-config-data\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.281263 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.291381 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-scripts\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.308347 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mk5v\" (UniqueName: \"kubernetes.io/projected/094675fb-839b-47fe-b1cd-dec2fafbf906-kube-api-access-6mk5v\") pod \"nova-cell0-conductor-db-sync-z4vqv\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.377432 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.767751 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.899755 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerStarted","Data":"121662ab5a206dc3722006d43ef8fe9c8ed7b0a629f741a62e7e7b0daf041cf7"} Nov 23 20:26:33 crc kubenswrapper[4726]: W1123 20:26:33.901284 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod094675fb_839b_47fe_b1cd_dec2fafbf906.slice/crio-1374ad658e7cfbf675454663b814df2395996f877d365a3095eb8b76be3069ee WatchSource:0}: Error finding container 1374ad658e7cfbf675454663b814df2395996f877d365a3095eb8b76be3069ee: Status 404 returned error can't find the container with id 1374ad658e7cfbf675454663b814df2395996f877d365a3095eb8b76be3069ee Nov 23 20:26:33 crc kubenswrapper[4726]: I1123 20:26:33.909135 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z4vqv"] Nov 23 20:26:34 crc kubenswrapper[4726]: I1123 20:26:34.611356 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94debbed-92cd-4a0c-bc7a-f85692e013d0" path="/var/lib/kubelet/pods/94debbed-92cd-4a0c-bc7a-f85692e013d0/volumes" Nov 23 20:26:34 crc kubenswrapper[4726]: I1123 20:26:34.912098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" event={"ID":"094675fb-839b-47fe-b1cd-dec2fafbf906","Type":"ContainerStarted","Data":"1374ad658e7cfbf675454663b814df2395996f877d365a3095eb8b76be3069ee"} Nov 23 20:26:34 crc kubenswrapper[4726]: I1123 20:26:34.915487 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerStarted","Data":"50df4a1002659e2dd9fd89082fbcaf496413a45f6d210d6b56a05c3b9f1b85f7"} Nov 23 20:26:35 crc kubenswrapper[4726]: I1123 20:26:35.941334 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerStarted","Data":"bfce87790b8cb5a2731725696a805abac9f05e64438ad730fe8fdbf8a3c6b328"} Nov 23 20:26:35 crc kubenswrapper[4726]: I1123 20:26:35.941854 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerStarted","Data":"e3a32e9ff80d94d1342b7fb55687f0a250dd8b3db6e5a507ca76b4eca4c67d5f"} Nov 23 20:26:39 crc kubenswrapper[4726]: I1123 20:26:39.043379 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:26:39 crc kubenswrapper[4726]: I1123 20:26:39.043795 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:26:42 crc kubenswrapper[4726]: I1123 20:26:42.047797 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" event={"ID":"094675fb-839b-47fe-b1cd-dec2fafbf906","Type":"ContainerStarted","Data":"5882b6fe29c83f14108e1f4ee991c9038b632d758d4db3ca9b62ef0cf8deafa4"} Nov 23 20:26:42 crc kubenswrapper[4726]: I1123 20:26:42.055703 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerStarted","Data":"622d7f8e8f3ba17b76dc24fad3baba0a26d32014c8b0e1a6a0f7d77b5484c73a"} Nov 23 20:26:42 crc kubenswrapper[4726]: I1123 20:26:42.056192 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 20:26:42 crc kubenswrapper[4726]: I1123 20:26:42.083716 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" podStartSLOduration=1.8339337759999998 podStartE2EDuration="9.083687176s" podCreationTimestamp="2025-11-23 20:26:33 +0000 UTC" firstStartedPulling="2025-11-23 20:26:33.90457879 +0000 UTC m=+1102.053619736" lastFinishedPulling="2025-11-23 20:26:41.15433218 +0000 UTC m=+1109.303373136" observedRunningTime="2025-11-23 20:26:42.07184854 +0000 UTC m=+1110.220889516" watchObservedRunningTime="2025-11-23 20:26:42.083687176 +0000 UTC m=+1110.232728172" Nov 23 20:26:52 crc kubenswrapper[4726]: I1123 20:26:52.162512 4726 generic.go:334] "Generic (PLEG): container finished" podID="094675fb-839b-47fe-b1cd-dec2fafbf906" containerID="5882b6fe29c83f14108e1f4ee991c9038b632d758d4db3ca9b62ef0cf8deafa4" exitCode=0 Nov 23 20:26:52 crc kubenswrapper[4726]: I1123 20:26:52.162683 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" event={"ID":"094675fb-839b-47fe-b1cd-dec2fafbf906","Type":"ContainerDied","Data":"5882b6fe29c83f14108e1f4ee991c9038b632d758d4db3ca9b62ef0cf8deafa4"} Nov 23 20:26:52 crc kubenswrapper[4726]: I1123 20:26:52.180784 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=12.807232931 podStartE2EDuration="20.18075795s" podCreationTimestamp="2025-11-23 20:26:32 +0000 UTC" firstStartedPulling="2025-11-23 20:26:33.779521357 +0000 UTC m=+1101.928562313" lastFinishedPulling="2025-11-23 20:26:41.153046336 +0000 UTC m=+1109.302087332" observedRunningTime="2025-11-23 20:26:42.110350916 +0000 UTC m=+1110.259391932" watchObservedRunningTime="2025-11-23 20:26:52.18075795 +0000 UTC m=+1120.329798916" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.590704 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.757192 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-combined-ca-bundle\") pod \"094675fb-839b-47fe-b1cd-dec2fafbf906\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.757298 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-scripts\") pod \"094675fb-839b-47fe-b1cd-dec2fafbf906\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.757476 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mk5v\" (UniqueName: \"kubernetes.io/projected/094675fb-839b-47fe-b1cd-dec2fafbf906-kube-api-access-6mk5v\") pod \"094675fb-839b-47fe-b1cd-dec2fafbf906\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.757595 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-config-data\") pod \"094675fb-839b-47fe-b1cd-dec2fafbf906\" (UID: \"094675fb-839b-47fe-b1cd-dec2fafbf906\") " Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.767780 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/094675fb-839b-47fe-b1cd-dec2fafbf906-kube-api-access-6mk5v" (OuterVolumeSpecName: "kube-api-access-6mk5v") pod "094675fb-839b-47fe-b1cd-dec2fafbf906" (UID: "094675fb-839b-47fe-b1cd-dec2fafbf906"). InnerVolumeSpecName "kube-api-access-6mk5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.768559 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-scripts" (OuterVolumeSpecName: "scripts") pod "094675fb-839b-47fe-b1cd-dec2fafbf906" (UID: "094675fb-839b-47fe-b1cd-dec2fafbf906"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.785253 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "094675fb-839b-47fe-b1cd-dec2fafbf906" (UID: "094675fb-839b-47fe-b1cd-dec2fafbf906"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.805386 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-config-data" (OuterVolumeSpecName: "config-data") pod "094675fb-839b-47fe-b1cd-dec2fafbf906" (UID: "094675fb-839b-47fe-b1cd-dec2fafbf906"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.860624 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.860675 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.860690 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mk5v\" (UniqueName: \"kubernetes.io/projected/094675fb-839b-47fe-b1cd-dec2fafbf906-kube-api-access-6mk5v\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:53 crc kubenswrapper[4726]: I1123 20:26:53.860703 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094675fb-839b-47fe-b1cd-dec2fafbf906-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.187229 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" event={"ID":"094675fb-839b-47fe-b1cd-dec2fafbf906","Type":"ContainerDied","Data":"1374ad658e7cfbf675454663b814df2395996f877d365a3095eb8b76be3069ee"} Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.187271 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1374ad658e7cfbf675454663b814df2395996f877d365a3095eb8b76be3069ee" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.187334 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z4vqv" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.347204 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 20:26:54 crc kubenswrapper[4726]: E1123 20:26:54.347713 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094675fb-839b-47fe-b1cd-dec2fafbf906" containerName="nova-cell0-conductor-db-sync" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.347734 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="094675fb-839b-47fe-b1cd-dec2fafbf906" containerName="nova-cell0-conductor-db-sync" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.348033 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="094675fb-839b-47fe-b1cd-dec2fafbf906" containerName="nova-cell0-conductor-db-sync" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.348779 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.352291 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-w42x6" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.352545 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.365258 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.472415 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk62z\" (UniqueName: \"kubernetes.io/projected/a327e21d-682a-4006-9ef1-b6a07276fa3d-kube-api-access-hk62z\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.472476 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a327e21d-682a-4006-9ef1-b6a07276fa3d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.472526 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a327e21d-682a-4006-9ef1-b6a07276fa3d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.574334 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a327e21d-682a-4006-9ef1-b6a07276fa3d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.574822 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk62z\" (UniqueName: \"kubernetes.io/projected/a327e21d-682a-4006-9ef1-b6a07276fa3d-kube-api-access-hk62z\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.575053 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a327e21d-682a-4006-9ef1-b6a07276fa3d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.580511 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a327e21d-682a-4006-9ef1-b6a07276fa3d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.581668 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a327e21d-682a-4006-9ef1-b6a07276fa3d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.598506 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk62z\" (UniqueName: \"kubernetes.io/projected/a327e21d-682a-4006-9ef1-b6a07276fa3d-kube-api-access-hk62z\") pod \"nova-cell0-conductor-0\" (UID: \"a327e21d-682a-4006-9ef1-b6a07276fa3d\") " pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:54 crc kubenswrapper[4726]: I1123 20:26:54.664560 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:55 crc kubenswrapper[4726]: I1123 20:26:55.141744 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 23 20:26:55 crc kubenswrapper[4726]: I1123 20:26:55.206087 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a327e21d-682a-4006-9ef1-b6a07276fa3d","Type":"ContainerStarted","Data":"a28ce6256ad203547f288f555c3bacd5e6dcf133934b9c99d18980897fba6f16"} Nov 23 20:26:56 crc kubenswrapper[4726]: I1123 20:26:56.218509 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a327e21d-682a-4006-9ef1-b6a07276fa3d","Type":"ContainerStarted","Data":"56045995cf8ffcdc22a58b88a582ff39a394ed8ad8503048b39ec1ce871a9916"} Nov 23 20:26:56 crc kubenswrapper[4726]: I1123 20:26:56.219043 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 23 20:26:56 crc kubenswrapper[4726]: I1123 20:26:56.253768 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.253739717 podStartE2EDuration="2.253739717s" podCreationTimestamp="2025-11-23 20:26:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:26:56.24675032 +0000 UTC m=+1124.395791296" watchObservedRunningTime="2025-11-23 20:26:56.253739717 +0000 UTC m=+1124.402780713" Nov 23 20:27:03 crc kubenswrapper[4726]: I1123 20:27:03.283465 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 20:27:04 crc kubenswrapper[4726]: I1123 20:27:04.695489 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.312823 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-dqb9v"] Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.313929 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.323660 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.323715 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.353913 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-dqb9v"] Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.423339 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-config-data\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.423427 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82txh\" (UniqueName: \"kubernetes.io/projected/80fe916a-801a-4d1e-a9b1-fb819406e39f-kube-api-access-82txh\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.423743 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.423959 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-scripts\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.526141 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.526228 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-scripts\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.526265 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-config-data\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.526305 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82txh\" (UniqueName: \"kubernetes.io/projected/80fe916a-801a-4d1e-a9b1-fb819406e39f-kube-api-access-82txh\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.533641 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-config-data\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.533739 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-scripts\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.534562 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.575485 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82txh\" (UniqueName: \"kubernetes.io/projected/80fe916a-801a-4d1e-a9b1-fb819406e39f-kube-api-access-82txh\") pod \"nova-cell0-cell-mapping-dqb9v\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.633232 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.762951 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.767371 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.790171 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.801544 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.862046 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.868340 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.877118 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.941858 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw4cp\" (UniqueName: \"kubernetes.io/projected/847e98eb-74a9-4936-990a-881fbc0b2271-kube-api-access-qw4cp\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.941923 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.941976 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-config-data\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:05 crc kubenswrapper[4726]: I1123 20:27:05.970579 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.030151 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.031643 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.049120 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-config-data\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.049251 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b668ba-4608-4fa3-8342-98f1af7f5263-logs\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.049273 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.049354 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdwt8\" (UniqueName: \"kubernetes.io/projected/99b668ba-4608-4fa3-8342-98f1af7f5263-kube-api-access-zdwt8\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.049378 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw4cp\" (UniqueName: \"kubernetes.io/projected/847e98eb-74a9-4936-990a-881fbc0b2271-kube-api-access-qw4cp\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.049396 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-config-data\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.049415 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.050283 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.060348 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.072388 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-config-data\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.072841 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.108471 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw4cp\" (UniqueName: \"kubernetes.io/projected/847e98eb-74a9-4936-990a-881fbc0b2271-kube-api-access-qw4cp\") pod \"nova-scheduler-0\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.121501 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.156829 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcrs8\" (UniqueName: \"kubernetes.io/projected/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-kube-api-access-qcrs8\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.157746 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.177286 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.192070 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.209264 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.209349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-config-data\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.214672 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdwt8\" (UniqueName: \"kubernetes.io/projected/99b668ba-4608-4fa3-8342-98f1af7f5263-kube-api-access-zdwt8\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.214973 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-config-data\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.214999 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.215219 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-logs\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.215360 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b668ba-4608-4fa3-8342-98f1af7f5263-logs\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.215388 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.218215 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b668ba-4608-4fa3-8342-98f1af7f5263-logs\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.225602 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-config-data\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.230549 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.245529 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdwt8\" (UniqueName: \"kubernetes.io/projected/99b668ba-4608-4fa3-8342-98f1af7f5263-kube-api-access-zdwt8\") pod \"nova-api-0\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.352780 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-logs\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.353168 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcrs8\" (UniqueName: \"kubernetes.io/projected/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-kube-api-access-qcrs8\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.353282 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-config-data\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.353372 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.356318 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.356659 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.358184 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrrpv\" (UniqueName: \"kubernetes.io/projected/d771b2fa-fa53-48f5-ae54-112d535e04a3-kube-api-access-lrrpv\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.359003 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-logs\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.355034 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-gfjd8"] Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.363495 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.379710 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.397204 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-config-data\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.401271 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-gfjd8"] Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.408533 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcrs8\" (UniqueName: \"kubernetes.io/projected/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-kube-api-access-qcrs8\") pod \"nova-metadata-0\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.460561 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-nb\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.460705 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-config\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.460755 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.460793 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.460820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-sb\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.460882 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-dns-svc\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.460910 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrrpv\" (UniqueName: \"kubernetes.io/projected/d771b2fa-fa53-48f5-ae54-112d535e04a3-kube-api-access-lrrpv\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.460958 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzkc7\" (UniqueName: \"kubernetes.io/projected/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-kube-api-access-vzkc7\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.472040 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.472521 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.478666 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.505531 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrrpv\" (UniqueName: \"kubernetes.io/projected/d771b2fa-fa53-48f5-ae54-112d535e04a3-kube-api-access-lrrpv\") pod \"nova-cell1-novncproxy-0\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.506063 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.545369 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.563576 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-config\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.563685 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-sb\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.563746 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-dns-svc\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.563818 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzkc7\" (UniqueName: \"kubernetes.io/projected/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-kube-api-access-vzkc7\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.564028 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-nb\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.566011 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-dns-svc\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.566134 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-config\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.566509 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-sb\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.566653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-nb\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.596775 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzkc7\" (UniqueName: \"kubernetes.io/projected/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-kube-api-access-vzkc7\") pod \"dnsmasq-dns-69494d9f89-gfjd8\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.660918 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-dqb9v"] Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.707418 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:06 crc kubenswrapper[4726]: I1123 20:27:06.894447 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:06 crc kubenswrapper[4726]: W1123 20:27:06.923235 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod847e98eb_74a9_4936_990a_881fbc0b2271.slice/crio-d55ee7cc8683d7b57cf5d0f90bb115847f26d36edaa3d173f78ca3d2ef2b5d38 WatchSource:0}: Error finding container d55ee7cc8683d7b57cf5d0f90bb115847f26d36edaa3d173f78ca3d2ef2b5d38: Status 404 returned error can't find the container with id d55ee7cc8683d7b57cf5d0f90bb115847f26d36edaa3d173f78ca3d2ef2b5d38 Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.222109 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:07 crc kubenswrapper[4726]: W1123 20:27:07.224020 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ae2ce8d_ca20_4cd2_810b_a5b24ab623c8.slice/crio-ff72b0e8725bef3c65f26e81ad937c44537113d6211535408f9500d933c25450 WatchSource:0}: Error finding container ff72b0e8725bef3c65f26e81ad937c44537113d6211535408f9500d933c25450: Status 404 returned error can't find the container with id ff72b0e8725bef3c65f26e81ad937c44537113d6211535408f9500d933c25450 Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.382914 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:07 crc kubenswrapper[4726]: W1123 20:27:07.385450 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd771b2fa_fa53_48f5_ae54_112d535e04a3.slice/crio-21b5277fe418174dbae3716a0c141ae33d3e360733c0e7d055826649d409aa73 WatchSource:0}: Error finding container 21b5277fe418174dbae3716a0c141ae33d3e360733c0e7d055826649d409aa73: Status 404 returned error can't find the container with id 21b5277fe418174dbae3716a0c141ae33d3e360733c0e7d055826649d409aa73 Nov 23 20:27:07 crc kubenswrapper[4726]: W1123 20:27:07.389973 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99b668ba_4608_4fa3_8342_98f1af7f5263.slice/crio-6edda3020a88e01aca489c2d2d8cab7d64a0170264e3a2cf854d19ac3bd9cfb7 WatchSource:0}: Error finding container 6edda3020a88e01aca489c2d2d8cab7d64a0170264e3a2cf854d19ac3bd9cfb7: Status 404 returned error can't find the container with id 6edda3020a88e01aca489c2d2d8cab7d64a0170264e3a2cf854d19ac3bd9cfb7 Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.402751 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.419635 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99b668ba-4608-4fa3-8342-98f1af7f5263","Type":"ContainerStarted","Data":"6edda3020a88e01aca489c2d2d8cab7d64a0170264e3a2cf854d19ac3bd9cfb7"} Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.426555 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d771b2fa-fa53-48f5-ae54-112d535e04a3","Type":"ContainerStarted","Data":"21b5277fe418174dbae3716a0c141ae33d3e360733c0e7d055826649d409aa73"} Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.443135 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8","Type":"ContainerStarted","Data":"ff72b0e8725bef3c65f26e81ad937c44537113d6211535408f9500d933c25450"} Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.456953 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dqb9v" event={"ID":"80fe916a-801a-4d1e-a9b1-fb819406e39f","Type":"ContainerStarted","Data":"f49aefbb1bd6feb8f32dec14ca09ac31201d7769cec209ae7c463464c7edf83b"} Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.457180 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dqb9v" event={"ID":"80fe916a-801a-4d1e-a9b1-fb819406e39f","Type":"ContainerStarted","Data":"6164cb80a06849d6fc659cfeaf1dda58bc577c777f5675f811bf82149d7f844c"} Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.464540 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qps7j"] Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.466273 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.475240 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.475538 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.480134 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"847e98eb-74a9-4936-990a-881fbc0b2271","Type":"ContainerStarted","Data":"d55ee7cc8683d7b57cf5d0f90bb115847f26d36edaa3d173f78ca3d2ef2b5d38"} Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.503511 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qps7j"] Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.504752 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-config-data\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.504803 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rddzq\" (UniqueName: \"kubernetes.io/projected/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-kube-api-access-rddzq\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.504955 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-scripts\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.505002 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.520313 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-dqb9v" podStartSLOduration=2.520285066 podStartE2EDuration="2.520285066s" podCreationTimestamp="2025-11-23 20:27:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:07.499261316 +0000 UTC m=+1135.648302272" watchObservedRunningTime="2025-11-23 20:27:07.520285066 +0000 UTC m=+1135.669326022" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.608144 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-scripts\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.608221 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.608255 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-config-data\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.608278 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rddzq\" (UniqueName: \"kubernetes.io/projected/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-kube-api-access-rddzq\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.649460 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-scripts\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.659500 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-config-data\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.676414 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rddzq\" (UniqueName: \"kubernetes.io/projected/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-kube-api-access-rddzq\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.685528 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qps7j\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.792343 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:07 crc kubenswrapper[4726]: I1123 20:27:07.918091 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-gfjd8"] Nov 23 20:27:07 crc kubenswrapper[4726]: W1123 20:27:07.922097 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3ab62d2_07f5_4a6b_95aa_e0f52f9b0613.slice/crio-49db783d1234e2fa407df204f7e99d6733f28b4307fe5e1203b55cfce525bdee WatchSource:0}: Error finding container 49db783d1234e2fa407df204f7e99d6733f28b4307fe5e1203b55cfce525bdee: Status 404 returned error can't find the container with id 49db783d1234e2fa407df204f7e99d6733f28b4307fe5e1203b55cfce525bdee Nov 23 20:27:08 crc kubenswrapper[4726]: I1123 20:27:08.515395 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" event={"ID":"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613","Type":"ContainerStarted","Data":"93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494"} Nov 23 20:27:08 crc kubenswrapper[4726]: I1123 20:27:08.515842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" event={"ID":"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613","Type":"ContainerStarted","Data":"49db783d1234e2fa407df204f7e99d6733f28b4307fe5e1203b55cfce525bdee"} Nov 23 20:27:08 crc kubenswrapper[4726]: I1123 20:27:08.614635 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qps7j"] Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.043487 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.043760 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.043804 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.044573 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93d2133e107e131190f68c969bc423a889dae2dbf2757948997b4402e77bc885"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.044623 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://93d2133e107e131190f68c969bc423a889dae2dbf2757948997b4402e77bc885" gracePeriod=600 Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.535590 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="93d2133e107e131190f68c969bc423a889dae2dbf2757948997b4402e77bc885" exitCode=0 Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.535722 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"93d2133e107e131190f68c969bc423a889dae2dbf2757948997b4402e77bc885"} Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.535761 4726 scope.go:117] "RemoveContainer" containerID="e0e5405a0e8a49ddd9351d6065e06af6eb07dcd086934b09caf4e017ae67cf3f" Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.538230 4726 generic.go:334] "Generic (PLEG): container finished" podID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" containerID="93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494" exitCode=0 Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.538305 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" event={"ID":"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613","Type":"ContainerDied","Data":"93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494"} Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.540156 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qps7j" event={"ID":"8a8bd6c7-c313-475a-95ce-d95fb659ebaf","Type":"ContainerStarted","Data":"e085e9ab6c370835a2062856c60669b7989070212f1dacc717ed01fb1746eace"} Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.540191 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qps7j" event={"ID":"8a8bd6c7-c313-475a-95ce-d95fb659ebaf","Type":"ContainerStarted","Data":"7a70a63870d5c0ac0bd6e95a4e91ecd33143fc42916e8b586e7d1eafd74f5fd6"} Nov 23 20:27:09 crc kubenswrapper[4726]: I1123 20:27:09.574209 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-qps7j" podStartSLOduration=2.574187898 podStartE2EDuration="2.574187898s" podCreationTimestamp="2025-11-23 20:27:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:09.571591759 +0000 UTC m=+1137.720632725" watchObservedRunningTime="2025-11-23 20:27:09.574187898 +0000 UTC m=+1137.723228854" Nov 23 20:27:10 crc kubenswrapper[4726]: I1123 20:27:10.766073 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:10 crc kubenswrapper[4726]: I1123 20:27:10.777115 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:27:10 crc kubenswrapper[4726]: I1123 20:27:10.777362 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="91c39827-8843-4434-8e7f-1ff5be6d9c24" containerName="kube-state-metrics" containerID="cri-o://afda28590e38150743d2f7734882f8429d718e8fbd2782bf68a9341cdabe347a" gracePeriod=30 Nov 23 20:27:10 crc kubenswrapper[4726]: I1123 20:27:10.797111 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:11 crc kubenswrapper[4726]: I1123 20:27:11.608629 4726 generic.go:334] "Generic (PLEG): container finished" podID="91c39827-8843-4434-8e7f-1ff5be6d9c24" containerID="afda28590e38150743d2f7734882f8429d718e8fbd2782bf68a9341cdabe347a" exitCode=2 Nov 23 20:27:11 crc kubenswrapper[4726]: I1123 20:27:11.609209 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"91c39827-8843-4434-8e7f-1ff5be6d9c24","Type":"ContainerDied","Data":"afda28590e38150743d2f7734882f8429d718e8fbd2782bf68a9341cdabe347a"} Nov 23 20:27:11 crc kubenswrapper[4726]: I1123 20:27:11.822897 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 20:27:11 crc kubenswrapper[4726]: I1123 20:27:11.932557 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbb2j\" (UniqueName: \"kubernetes.io/projected/91c39827-8843-4434-8e7f-1ff5be6d9c24-kube-api-access-fbb2j\") pod \"91c39827-8843-4434-8e7f-1ff5be6d9c24\" (UID: \"91c39827-8843-4434-8e7f-1ff5be6d9c24\") " Nov 23 20:27:11 crc kubenswrapper[4726]: I1123 20:27:11.988321 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91c39827-8843-4434-8e7f-1ff5be6d9c24-kube-api-access-fbb2j" (OuterVolumeSpecName: "kube-api-access-fbb2j") pod "91c39827-8843-4434-8e7f-1ff5be6d9c24" (UID: "91c39827-8843-4434-8e7f-1ff5be6d9c24"). InnerVolumeSpecName "kube-api-access-fbb2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.040958 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbb2j\" (UniqueName: \"kubernetes.io/projected/91c39827-8843-4434-8e7f-1ff5be6d9c24-kube-api-access-fbb2j\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.632332 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d771b2fa-fa53-48f5-ae54-112d535e04a3","Type":"ContainerStarted","Data":"2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.632419 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="d771b2fa-fa53-48f5-ae54-112d535e04a3" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f" gracePeriod=30 Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.639171 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8","Type":"ContainerStarted","Data":"e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.639253 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8","Type":"ContainerStarted","Data":"1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.639431 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerName="nova-metadata-log" containerID="cri-o://1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491" gracePeriod=30 Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.639836 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerName="nova-metadata-metadata" containerID="cri-o://e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753" gracePeriod=30 Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.656337 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"847e98eb-74a9-4936-990a-881fbc0b2271","Type":"ContainerStarted","Data":"6fdcd167e72753d7df9099060cd79c81fe26cdd50cbc71f1fea8960e030d3c08"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.671248 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"91c39827-8843-4434-8e7f-1ff5be6d9c24","Type":"ContainerDied","Data":"a80199e7223b40bfb77f2ae14fce97328f16719b49cd6b3c78a04c2786f6b8b6"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.671304 4726 scope.go:117] "RemoveContainer" containerID="afda28590e38150743d2f7734882f8429d718e8fbd2782bf68a9341cdabe347a" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.671467 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.681225 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99b668ba-4608-4fa3-8342-98f1af7f5263","Type":"ContainerStarted","Data":"4284d64b096fcdd9d651c217914f79124327dd0d66f13378e0c7718153fae473"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.681268 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99b668ba-4608-4fa3-8342-98f1af7f5263","Type":"ContainerStarted","Data":"8da6a9027874c9ce99e53b9df7d1f66f74dd84c747f9445e3d8b63049c9e7513"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.692319 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"828799a4d998638471568a30864015defe9240aef8698bc3b256f460fe153d23"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.699756 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" event={"ID":"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613","Type":"ContainerStarted","Data":"b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102"} Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.700681 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.748056 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.777772 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.591322877 podStartE2EDuration="7.777753977s" podCreationTimestamp="2025-11-23 20:27:05 +0000 UTC" firstStartedPulling="2025-11-23 20:27:07.227200596 +0000 UTC m=+1135.376241552" lastFinishedPulling="2025-11-23 20:27:11.413631696 +0000 UTC m=+1139.562672652" observedRunningTime="2025-11-23 20:27:12.737811063 +0000 UTC m=+1140.886852019" watchObservedRunningTime="2025-11-23 20:27:12.777753977 +0000 UTC m=+1140.926794933" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.777972 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.838501 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:27:12 crc kubenswrapper[4726]: E1123 20:27:12.838915 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91c39827-8843-4434-8e7f-1ff5be6d9c24" containerName="kube-state-metrics" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.838927 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="91c39827-8843-4434-8e7f-1ff5be6d9c24" containerName="kube-state-metrics" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.839116 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="91c39827-8843-4434-8e7f-1ff5be6d9c24" containerName="kube-state-metrics" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.839738 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.841928 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.842117 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.876507 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.898709 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.44247278 podStartE2EDuration="7.898689989s" podCreationTimestamp="2025-11-23 20:27:05 +0000 UTC" firstStartedPulling="2025-11-23 20:27:06.946769043 +0000 UTC m=+1135.095809999" lastFinishedPulling="2025-11-23 20:27:11.402986252 +0000 UTC m=+1139.552027208" observedRunningTime="2025-11-23 20:27:12.807226382 +0000 UTC m=+1140.956267348" watchObservedRunningTime="2025-11-23 20:27:12.898689989 +0000 UTC m=+1141.047730945" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.920142 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.9016964769999998 podStartE2EDuration="6.92012546s" podCreationTimestamp="2025-11-23 20:27:06 +0000 UTC" firstStartedPulling="2025-11-23 20:27:07.395219063 +0000 UTC m=+1135.544260009" lastFinishedPulling="2025-11-23 20:27:11.413648036 +0000 UTC m=+1139.562688992" observedRunningTime="2025-11-23 20:27:12.862292379 +0000 UTC m=+1141.011333335" watchObservedRunningTime="2025-11-23 20:27:12.92012546 +0000 UTC m=+1141.069166416" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.933336 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.916601483 podStartE2EDuration="7.933319761s" podCreationTimestamp="2025-11-23 20:27:05 +0000 UTC" firstStartedPulling="2025-11-23 20:27:07.398148601 +0000 UTC m=+1135.547189557" lastFinishedPulling="2025-11-23 20:27:11.414866879 +0000 UTC m=+1139.563907835" observedRunningTime="2025-11-23 20:27:12.883825663 +0000 UTC m=+1141.032866629" watchObservedRunningTime="2025-11-23 20:27:12.933319761 +0000 UTC m=+1141.082360707" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.944157 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" podStartSLOduration=6.94413482 podStartE2EDuration="6.94413482s" podCreationTimestamp="2025-11-23 20:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:12.907308998 +0000 UTC m=+1141.056349964" watchObservedRunningTime="2025-11-23 20:27:12.94413482 +0000 UTC m=+1141.093175786" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.959600 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggtcl\" (UniqueName: \"kubernetes.io/projected/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-api-access-ggtcl\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.959651 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.959733 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:12 crc kubenswrapper[4726]: I1123 20:27:12.959801 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.061301 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggtcl\" (UniqueName: \"kubernetes.io/projected/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-api-access-ggtcl\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.061675 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.062231 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.062315 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.066769 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.067290 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.087137 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.090484 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggtcl\" (UniqueName: \"kubernetes.io/projected/71a1f62d-5ad3-450a-824a-bff237a8252e-kube-api-access-ggtcl\") pod \"kube-state-metrics-0\" (UID: \"71a1f62d-5ad3-450a-824a-bff237a8252e\") " pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.170337 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.170643 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="ceilometer-central-agent" containerID="cri-o://50df4a1002659e2dd9fd89082fbcaf496413a45f6d210d6b56a05c3b9f1b85f7" gracePeriod=30 Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.171181 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="proxy-httpd" containerID="cri-o://622d7f8e8f3ba17b76dc24fad3baba0a26d32014c8b0e1a6a0f7d77b5484c73a" gracePeriod=30 Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.171269 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="sg-core" containerID="cri-o://bfce87790b8cb5a2731725696a805abac9f05e64438ad730fe8fdbf8a3c6b328" gracePeriod=30 Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.171330 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="ceilometer-notification-agent" containerID="cri-o://e3a32e9ff80d94d1342b7fb55687f0a250dd8b3db6e5a507ca76b4eca4c67d5f" gracePeriod=30 Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.172832 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.713744 4726 generic.go:334] "Generic (PLEG): container finished" podID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerID="1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491" exitCode=143 Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.714306 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8","Type":"ContainerDied","Data":"1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491"} Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.749708 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerID="622d7f8e8f3ba17b76dc24fad3baba0a26d32014c8b0e1a6a0f7d77b5484c73a" exitCode=0 Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.750091 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerID="bfce87790b8cb5a2731725696a805abac9f05e64438ad730fe8fdbf8a3c6b328" exitCode=2 Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.750146 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerDied","Data":"622d7f8e8f3ba17b76dc24fad3baba0a26d32014c8b0e1a6a0f7d77b5484c73a"} Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.750174 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerDied","Data":"bfce87790b8cb5a2731725696a805abac9f05e64438ad730fe8fdbf8a3c6b328"} Nov 23 20:27:13 crc kubenswrapper[4726]: I1123 20:27:13.946146 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 23 20:27:14 crc kubenswrapper[4726]: I1123 20:27:14.599095 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91c39827-8843-4434-8e7f-1ff5be6d9c24" path="/var/lib/kubelet/pods/91c39827-8843-4434-8e7f-1ff5be6d9c24/volumes" Nov 23 20:27:14 crc kubenswrapper[4726]: I1123 20:27:14.768278 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerID="50df4a1002659e2dd9fd89082fbcaf496413a45f6d210d6b56a05c3b9f1b85f7" exitCode=0 Nov 23 20:27:14 crc kubenswrapper[4726]: I1123 20:27:14.768383 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerDied","Data":"50df4a1002659e2dd9fd89082fbcaf496413a45f6d210d6b56a05c3b9f1b85f7"} Nov 23 20:27:14 crc kubenswrapper[4726]: I1123 20:27:14.770367 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"71a1f62d-5ad3-450a-824a-bff237a8252e","Type":"ContainerStarted","Data":"e7bf1124b03c9027076e91e31318e53f965bddbad608746066dfb9304c419d03"} Nov 23 20:27:14 crc kubenswrapper[4726]: I1123 20:27:14.770406 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"71a1f62d-5ad3-450a-824a-bff237a8252e","Type":"ContainerStarted","Data":"c1a0b9eda0f99145d9e4b845b2d732ff0dff8614409d1907e2ab63bb7508588a"} Nov 23 20:27:14 crc kubenswrapper[4726]: I1123 20:27:14.770467 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 23 20:27:14 crc kubenswrapper[4726]: I1123 20:27:14.787156 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.408122402 podStartE2EDuration="2.787129812s" podCreationTimestamp="2025-11-23 20:27:12 +0000 UTC" firstStartedPulling="2025-11-23 20:27:14.011990756 +0000 UTC m=+1142.161031712" lastFinishedPulling="2025-11-23 20:27:14.390998166 +0000 UTC m=+1142.540039122" observedRunningTime="2025-11-23 20:27:14.785753605 +0000 UTC m=+1142.934794561" watchObservedRunningTime="2025-11-23 20:27:14.787129812 +0000 UTC m=+1142.936170778" Nov 23 20:27:15 crc kubenswrapper[4726]: I1123 20:27:15.994369 4726 scope.go:117] "RemoveContainer" containerID="12dae09b369d39fc8283c67b6e3897ba96fb2c7b98ddbeb7a65c1d0160a048de" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.035763 4726 scope.go:117] "RemoveContainer" containerID="20a43ef31eaf6ee3fe2af6a7c38f5172cb85af4cd44a152dcd81a05fd95f8453" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.097781 4726 scope.go:117] "RemoveContainer" containerID="925ec625f921f425b67467e2d98c6a00ee4b102e305d3acc944deccadb15c128" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.121883 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.121948 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.173482 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.479554 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.479614 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.507760 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.507824 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.546546 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.710074 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.764149 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-775457b975-ppzwh"] Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.764448 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-775457b975-ppzwh" podUID="b5b50700-1688-4d3f-8508-30a069c33ac7" containerName="dnsmasq-dns" containerID="cri-o://db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8" gracePeriod=10 Nov 23 20:27:16 crc kubenswrapper[4726]: I1123 20:27:16.856024 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.402022 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.437284 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-nb\") pod \"b5b50700-1688-4d3f-8508-30a069c33ac7\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.437364 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-config\") pod \"b5b50700-1688-4d3f-8508-30a069c33ac7\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.437423 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-sb\") pod \"b5b50700-1688-4d3f-8508-30a069c33ac7\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.437465 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-dns-svc\") pod \"b5b50700-1688-4d3f-8508-30a069c33ac7\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.437498 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn227\" (UniqueName: \"kubernetes.io/projected/b5b50700-1688-4d3f-8508-30a069c33ac7-kube-api-access-vn227\") pod \"b5b50700-1688-4d3f-8508-30a069c33ac7\" (UID: \"b5b50700-1688-4d3f-8508-30a069c33ac7\") " Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.484202 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b50700-1688-4d3f-8508-30a069c33ac7-kube-api-access-vn227" (OuterVolumeSpecName: "kube-api-access-vn227") pod "b5b50700-1688-4d3f-8508-30a069c33ac7" (UID: "b5b50700-1688-4d3f-8508-30a069c33ac7"). InnerVolumeSpecName "kube-api-access-vn227". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.542184 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn227\" (UniqueName: \"kubernetes.io/projected/b5b50700-1688-4d3f-8508-30a069c33ac7-kube-api-access-vn227\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.580620 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5b50700-1688-4d3f-8508-30a069c33ac7" (UID: "b5b50700-1688-4d3f-8508-30a069c33ac7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.607069 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.170:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.607464 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.170:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.635494 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5b50700-1688-4d3f-8508-30a069c33ac7" (UID: "b5b50700-1688-4d3f-8508-30a069c33ac7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.638929 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5b50700-1688-4d3f-8508-30a069c33ac7" (UID: "b5b50700-1688-4d3f-8508-30a069c33ac7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.648999 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.649416 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.649519 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.682443 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-config" (OuterVolumeSpecName: "config") pod "b5b50700-1688-4d3f-8508-30a069c33ac7" (UID: "b5b50700-1688-4d3f-8508-30a069c33ac7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.751770 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5b50700-1688-4d3f-8508-30a069c33ac7-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.875992 4726 generic.go:334] "Generic (PLEG): container finished" podID="b5b50700-1688-4d3f-8508-30a069c33ac7" containerID="db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8" exitCode=0 Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.876111 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-ppzwh" event={"ID":"b5b50700-1688-4d3f-8508-30a069c33ac7","Type":"ContainerDied","Data":"db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8"} Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.876155 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-775457b975-ppzwh" event={"ID":"b5b50700-1688-4d3f-8508-30a069c33ac7","Type":"ContainerDied","Data":"0521ca618178a57a352b7b7474947aa0b0fca73f3e03df21cf60df5ba443b89e"} Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.876180 4726 scope.go:117] "RemoveContainer" containerID="db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.876397 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-775457b975-ppzwh" Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.895576 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerID="e3a32e9ff80d94d1342b7fb55687f0a250dd8b3db6e5a507ca76b4eca4c67d5f" exitCode=0 Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.896031 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerDied","Data":"e3a32e9ff80d94d1342b7fb55687f0a250dd8b3db6e5a507ca76b4eca4c67d5f"} Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.923516 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-775457b975-ppzwh"] Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.936287 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-775457b975-ppzwh"] Nov 23 20:27:17 crc kubenswrapper[4726]: I1123 20:27:17.999070 4726 scope.go:117] "RemoveContainer" containerID="c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.067573 4726 scope.go:117] "RemoveContainer" containerID="db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8" Nov 23 20:27:18 crc kubenswrapper[4726]: E1123 20:27:18.070598 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8\": container with ID starting with db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8 not found: ID does not exist" containerID="db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.070643 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8"} err="failed to get container status \"db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8\": rpc error: code = NotFound desc = could not find container \"db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8\": container with ID starting with db8bcde15916df5d44e54afa72b159eed9af1171c8d93666b654b3e73fdd11d8 not found: ID does not exist" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.070671 4726 scope.go:117] "RemoveContainer" containerID="c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5" Nov 23 20:27:18 crc kubenswrapper[4726]: E1123 20:27:18.091006 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5\": container with ID starting with c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5 not found: ID does not exist" containerID="c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.091087 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5"} err="failed to get container status \"c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5\": rpc error: code = NotFound desc = could not find container \"c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5\": container with ID starting with c988f121b5b3d86acd72db3a1fe467b0336a11d2b20710e503c7f823ed5c07b5 not found: ID does not exist" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.275262 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.370829 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srf4v\" (UniqueName: \"kubernetes.io/projected/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-kube-api-access-srf4v\") pod \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.370929 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-combined-ca-bundle\") pod \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.370973 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-sg-core-conf-yaml\") pod \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.371028 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-run-httpd\") pod \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.371052 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-config-data\") pod \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.371086 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-scripts\") pod \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.371122 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-log-httpd\") pod \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\" (UID: \"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6\") " Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.371801 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" (UID: "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.372024 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.376134 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" (UID: "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.392002 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-kube-api-access-srf4v" (OuterVolumeSpecName: "kube-api-access-srf4v") pod "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" (UID: "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6"). InnerVolumeSpecName "kube-api-access-srf4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.427239 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-scripts" (OuterVolumeSpecName: "scripts") pod "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" (UID: "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.473756 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.473813 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.473854 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srf4v\" (UniqueName: \"kubernetes.io/projected/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-kube-api-access-srf4v\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.543065 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" (UID: "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.544759 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" (UID: "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.576465 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.576497 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.589987 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-config-data" (OuterVolumeSpecName: "config-data") pod "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" (UID: "8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.600350 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b50700-1688-4d3f-8508-30a069c33ac7" path="/var/lib/kubelet/pods/b5b50700-1688-4d3f-8508-30a069c33ac7/volumes" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.678051 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.908392 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6","Type":"ContainerDied","Data":"121662ab5a206dc3722006d43ef8fe9c8ed7b0a629f741a62e7e7b0daf041cf7"} Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.908440 4726 scope.go:117] "RemoveContainer" containerID="622d7f8e8f3ba17b76dc24fad3baba0a26d32014c8b0e1a6a0f7d77b5484c73a" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.908563 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.944289 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.955780 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.963651 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:18 crc kubenswrapper[4726]: E1123 20:27:18.964053 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b50700-1688-4d3f-8508-30a069c33ac7" containerName="init" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.964069 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b50700-1688-4d3f-8508-30a069c33ac7" containerName="init" Nov 23 20:27:18 crc kubenswrapper[4726]: E1123 20:27:18.964078 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="ceilometer-central-agent" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.964083 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="ceilometer-central-agent" Nov 23 20:27:18 crc kubenswrapper[4726]: E1123 20:27:18.964098 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="proxy-httpd" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.964104 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="proxy-httpd" Nov 23 20:27:18 crc kubenswrapper[4726]: E1123 20:27:18.964111 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="ceilometer-notification-agent" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.964118 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="ceilometer-notification-agent" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.967205 4726 scope.go:117] "RemoveContainer" containerID="bfce87790b8cb5a2731725696a805abac9f05e64438ad730fe8fdbf8a3c6b328" Nov 23 20:27:18 crc kubenswrapper[4726]: E1123 20:27:18.967943 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="sg-core" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.967970 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="sg-core" Nov 23 20:27:18 crc kubenswrapper[4726]: E1123 20:27:18.967989 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b50700-1688-4d3f-8508-30a069c33ac7" containerName="dnsmasq-dns" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.967996 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b50700-1688-4d3f-8508-30a069c33ac7" containerName="dnsmasq-dns" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.968288 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="sg-core" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.968309 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b50700-1688-4d3f-8508-30a069c33ac7" containerName="dnsmasq-dns" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.968327 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="proxy-httpd" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.968346 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="ceilometer-notification-agent" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.968355 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" containerName="ceilometer-central-agent" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.970069 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.980925 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 20:27:18 crc kubenswrapper[4726]: I1123 20:27:18.981197 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.012354 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.035129 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.079467 4726 scope.go:117] "RemoveContainer" containerID="e3a32e9ff80d94d1342b7fb55687f0a250dd8b3db6e5a507ca76b4eca4c67d5f" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.086694 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-config-data\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.086953 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.087141 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-run-httpd\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.087515 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-scripts\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.087620 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-log-httpd\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.087753 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.087925 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.088068 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rmw7\" (UniqueName: \"kubernetes.io/projected/c4cfd52a-db3c-4f20-89b3-a29921534365-kube-api-access-8rmw7\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.117083 4726 scope.go:117] "RemoveContainer" containerID="50df4a1002659e2dd9fd89082fbcaf496413a45f6d210d6b56a05c3b9f1b85f7" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.189668 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-config-data\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.189781 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.189805 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-run-httpd\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.189826 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-scripts\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.189843 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-log-httpd\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.189858 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.189910 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.189943 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rmw7\" (UniqueName: \"kubernetes.io/projected/c4cfd52a-db3c-4f20-89b3-a29921534365-kube-api-access-8rmw7\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.196539 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-log-httpd\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.197816 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-scripts\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.198144 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.198662 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-config-data\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.201317 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-run-httpd\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.207574 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.208236 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.216019 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rmw7\" (UniqueName: \"kubernetes.io/projected/c4cfd52a-db3c-4f20-89b3-a29921534365-kube-api-access-8rmw7\") pod \"ceilometer-0\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.337949 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:27:19 crc kubenswrapper[4726]: W1123 20:27:19.852176 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4cfd52a_db3c_4f20_89b3_a29921534365.slice/crio-ed290acc71239d479e7b960931260b18be2b04fdaa289a71bb992dac708cecf2 WatchSource:0}: Error finding container ed290acc71239d479e7b960931260b18be2b04fdaa289a71bb992dac708cecf2: Status 404 returned error can't find the container with id ed290acc71239d479e7b960931260b18be2b04fdaa289a71bb992dac708cecf2 Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.868220 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:19 crc kubenswrapper[4726]: I1123 20:27:19.919963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerStarted","Data":"ed290acc71239d479e7b960931260b18be2b04fdaa289a71bb992dac708cecf2"} Nov 23 20:27:20 crc kubenswrapper[4726]: I1123 20:27:20.605512 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6" path="/var/lib/kubelet/pods/8a02dda4-a8f4-4288-bbcf-dd5f5d9566a6/volumes" Nov 23 20:27:20 crc kubenswrapper[4726]: I1123 20:27:20.942969 4726 generic.go:334] "Generic (PLEG): container finished" podID="80fe916a-801a-4d1e-a9b1-fb819406e39f" containerID="f49aefbb1bd6feb8f32dec14ca09ac31201d7769cec209ae7c463464c7edf83b" exitCode=0 Nov 23 20:27:20 crc kubenswrapper[4726]: I1123 20:27:20.943046 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dqb9v" event={"ID":"80fe916a-801a-4d1e-a9b1-fb819406e39f","Type":"ContainerDied","Data":"f49aefbb1bd6feb8f32dec14ca09ac31201d7769cec209ae7c463464c7edf83b"} Nov 23 20:27:20 crc kubenswrapper[4726]: I1123 20:27:20.951683 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a8bd6c7-c313-475a-95ce-d95fb659ebaf" containerID="e085e9ab6c370835a2062856c60669b7989070212f1dacc717ed01fb1746eace" exitCode=0 Nov 23 20:27:20 crc kubenswrapper[4726]: I1123 20:27:20.951827 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qps7j" event={"ID":"8a8bd6c7-c313-475a-95ce-d95fb659ebaf","Type":"ContainerDied","Data":"e085e9ab6c370835a2062856c60669b7989070212f1dacc717ed01fb1746eace"} Nov 23 20:27:20 crc kubenswrapper[4726]: I1123 20:27:20.954200 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerStarted","Data":"d09d262f5944c3ae1d4472ee5dc1c1861c28e6609a121309e5e3cdf77b9026b8"} Nov 23 20:27:21 crc kubenswrapper[4726]: I1123 20:27:21.997804 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerStarted","Data":"b588c404918fc83ab4604909fba7fef861124df74fbfa6a0db6c600dd4f65366"} Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.496231 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.503436 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.680581 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-combined-ca-bundle\") pod \"80fe916a-801a-4d1e-a9b1-fb819406e39f\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.680979 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-scripts\") pod \"80fe916a-801a-4d1e-a9b1-fb819406e39f\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.681018 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-scripts\") pod \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.681057 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-config-data\") pod \"80fe916a-801a-4d1e-a9b1-fb819406e39f\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.681098 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-config-data\") pod \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.681550 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82txh\" (UniqueName: \"kubernetes.io/projected/80fe916a-801a-4d1e-a9b1-fb819406e39f-kube-api-access-82txh\") pod \"80fe916a-801a-4d1e-a9b1-fb819406e39f\" (UID: \"80fe916a-801a-4d1e-a9b1-fb819406e39f\") " Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.681611 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-combined-ca-bundle\") pod \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.681653 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rddzq\" (UniqueName: \"kubernetes.io/projected/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-kube-api-access-rddzq\") pod \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\" (UID: \"8a8bd6c7-c313-475a-95ce-d95fb659ebaf\") " Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.689691 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-kube-api-access-rddzq" (OuterVolumeSpecName: "kube-api-access-rddzq") pod "8a8bd6c7-c313-475a-95ce-d95fb659ebaf" (UID: "8a8bd6c7-c313-475a-95ce-d95fb659ebaf"). InnerVolumeSpecName "kube-api-access-rddzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.689758 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-scripts" (OuterVolumeSpecName: "scripts") pod "80fe916a-801a-4d1e-a9b1-fb819406e39f" (UID: "80fe916a-801a-4d1e-a9b1-fb819406e39f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.691608 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-scripts" (OuterVolumeSpecName: "scripts") pod "8a8bd6c7-c313-475a-95ce-d95fb659ebaf" (UID: "8a8bd6c7-c313-475a-95ce-d95fb659ebaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.691647 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80fe916a-801a-4d1e-a9b1-fb819406e39f-kube-api-access-82txh" (OuterVolumeSpecName: "kube-api-access-82txh") pod "80fe916a-801a-4d1e-a9b1-fb819406e39f" (UID: "80fe916a-801a-4d1e-a9b1-fb819406e39f"). InnerVolumeSpecName "kube-api-access-82txh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.723060 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a8bd6c7-c313-475a-95ce-d95fb659ebaf" (UID: "8a8bd6c7-c313-475a-95ce-d95fb659ebaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.723103 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-config-data" (OuterVolumeSpecName: "config-data") pod "8a8bd6c7-c313-475a-95ce-d95fb659ebaf" (UID: "8a8bd6c7-c313-475a-95ce-d95fb659ebaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.723176 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-config-data" (OuterVolumeSpecName: "config-data") pod "80fe916a-801a-4d1e-a9b1-fb819406e39f" (UID: "80fe916a-801a-4d1e-a9b1-fb819406e39f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.731287 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80fe916a-801a-4d1e-a9b1-fb819406e39f" (UID: "80fe916a-801a-4d1e-a9b1-fb819406e39f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.783681 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.783718 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rddzq\" (UniqueName: \"kubernetes.io/projected/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-kube-api-access-rddzq\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.783731 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.783741 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.783751 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.783761 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80fe916a-801a-4d1e-a9b1-fb819406e39f-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.783769 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8bd6c7-c313-475a-95ce-d95fb659ebaf-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:22 crc kubenswrapper[4726]: I1123 20:27:22.783780 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82txh\" (UniqueName: \"kubernetes.io/projected/80fe916a-801a-4d1e-a9b1-fb819406e39f-kube-api-access-82txh\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.009323 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-dqb9v" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.009348 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-dqb9v" event={"ID":"80fe916a-801a-4d1e-a9b1-fb819406e39f","Type":"ContainerDied","Data":"6164cb80a06849d6fc659cfeaf1dda58bc577c777f5675f811bf82149d7f844c"} Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.009403 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6164cb80a06849d6fc659cfeaf1dda58bc577c777f5675f811bf82149d7f844c" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.030425 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qps7j" event={"ID":"8a8bd6c7-c313-475a-95ce-d95fb659ebaf","Type":"ContainerDied","Data":"7a70a63870d5c0ac0bd6e95a4e91ecd33143fc42916e8b586e7d1eafd74f5fd6"} Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.030465 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a70a63870d5c0ac0bd6e95a4e91ecd33143fc42916e8b586e7d1eafd74f5fd6" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.030557 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qps7j" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.041698 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerStarted","Data":"8c21b267f39bb05298b4ae30e5df047d335ab6c459be6d35162237db33f72bc4"} Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.106016 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 20:27:23 crc kubenswrapper[4726]: E1123 20:27:23.108712 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fe916a-801a-4d1e-a9b1-fb819406e39f" containerName="nova-manage" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.108804 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fe916a-801a-4d1e-a9b1-fb819406e39f" containerName="nova-manage" Nov 23 20:27:23 crc kubenswrapper[4726]: E1123 20:27:23.108900 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a8bd6c7-c313-475a-95ce-d95fb659ebaf" containerName="nova-cell1-conductor-db-sync" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.108964 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a8bd6c7-c313-475a-95ce-d95fb659ebaf" containerName="nova-cell1-conductor-db-sync" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.109226 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a8bd6c7-c313-475a-95ce-d95fb659ebaf" containerName="nova-cell1-conductor-db-sync" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.109306 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="80fe916a-801a-4d1e-a9b1-fb819406e39f" containerName="nova-manage" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.110072 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.112366 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.124651 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.189007 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.201358 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d687a16-417f-4c97-b745-beba9b9429ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.201439 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d687a16-417f-4c97-b745-beba9b9429ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.201539 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkfgp\" (UniqueName: \"kubernetes.io/projected/5d687a16-417f-4c97-b745-beba9b9429ae-kube-api-access-zkfgp\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.284299 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.284562 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-log" containerID="cri-o://8da6a9027874c9ce99e53b9df7d1f66f74dd84c747f9445e3d8b63049c9e7513" gracePeriod=30 Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.285056 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-api" containerID="cri-o://4284d64b096fcdd9d651c217914f79124327dd0d66f13378e0c7718153fae473" gracePeriod=30 Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.305106 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d687a16-417f-4c97-b745-beba9b9429ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.305220 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d687a16-417f-4c97-b745-beba9b9429ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.305337 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkfgp\" (UniqueName: \"kubernetes.io/projected/5d687a16-417f-4c97-b745-beba9b9429ae-kube-api-access-zkfgp\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.306660 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.306905 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="847e98eb-74a9-4936-990a-881fbc0b2271" containerName="nova-scheduler-scheduler" containerID="cri-o://6fdcd167e72753d7df9099060cd79c81fe26cdd50cbc71f1fea8960e030d3c08" gracePeriod=30 Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.312967 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d687a16-417f-4c97-b745-beba9b9429ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.318653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d687a16-417f-4c97-b745-beba9b9429ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.372916 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkfgp\" (UniqueName: \"kubernetes.io/projected/5d687a16-417f-4c97-b745-beba9b9429ae-kube-api-access-zkfgp\") pod \"nova-cell1-conductor-0\" (UID: \"5d687a16-417f-4c97-b745-beba9b9429ae\") " pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.427846 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:23 crc kubenswrapper[4726]: I1123 20:27:23.959543 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 23 20:27:24 crc kubenswrapper[4726]: I1123 20:27:24.081354 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerStarted","Data":"c32e4b5a321337406904d31573f68e83c5cb223948120df2db90f9738e3c7048"} Nov 23 20:27:24 crc kubenswrapper[4726]: I1123 20:27:24.083061 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 20:27:24 crc kubenswrapper[4726]: I1123 20:27:24.089040 4726 generic.go:334] "Generic (PLEG): container finished" podID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerID="8da6a9027874c9ce99e53b9df7d1f66f74dd84c747f9445e3d8b63049c9e7513" exitCode=143 Nov 23 20:27:24 crc kubenswrapper[4726]: I1123 20:27:24.089131 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99b668ba-4608-4fa3-8342-98f1af7f5263","Type":"ContainerDied","Data":"8da6a9027874c9ce99e53b9df7d1f66f74dd84c747f9445e3d8b63049c9e7513"} Nov 23 20:27:24 crc kubenswrapper[4726]: I1123 20:27:24.098087 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5d687a16-417f-4c97-b745-beba9b9429ae","Type":"ContainerStarted","Data":"217c65ccae8daae5b082f126aecafb8a8d1ff27b24c7ce71157ba58cc676bf0a"} Nov 23 20:27:24 crc kubenswrapper[4726]: I1123 20:27:24.119742 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7122466530000002 podStartE2EDuration="6.119722135s" podCreationTimestamp="2025-11-23 20:27:18 +0000 UTC" firstStartedPulling="2025-11-23 20:27:19.854534558 +0000 UTC m=+1148.003575514" lastFinishedPulling="2025-11-23 20:27:23.26201005 +0000 UTC m=+1151.411050996" observedRunningTime="2025-11-23 20:27:24.110647593 +0000 UTC m=+1152.259688559" watchObservedRunningTime="2025-11-23 20:27:24.119722135 +0000 UTC m=+1152.268763091" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.107947 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5d687a16-417f-4c97-b745-beba9b9429ae","Type":"ContainerStarted","Data":"7ee2877b7f7c21b3cc8b8f54e63805821de5da108b9d8fbf5a5c6a1f218741bc"} Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.108435 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.111058 4726 generic.go:334] "Generic (PLEG): container finished" podID="847e98eb-74a9-4936-990a-881fbc0b2271" containerID="6fdcd167e72753d7df9099060cd79c81fe26cdd50cbc71f1fea8960e030d3c08" exitCode=0 Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.111403 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"847e98eb-74a9-4936-990a-881fbc0b2271","Type":"ContainerDied","Data":"6fdcd167e72753d7df9099060cd79c81fe26cdd50cbc71f1fea8960e030d3c08"} Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.111461 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"847e98eb-74a9-4936-990a-881fbc0b2271","Type":"ContainerDied","Data":"d55ee7cc8683d7b57cf5d0f90bb115847f26d36edaa3d173f78ca3d2ef2b5d38"} Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.111475 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d55ee7cc8683d7b57cf5d0f90bb115847f26d36edaa3d173f78ca3d2ef2b5d38" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.135057 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.135039121 podStartE2EDuration="2.135039121s" podCreationTimestamp="2025-11-23 20:27:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:25.133037318 +0000 UTC m=+1153.282078284" watchObservedRunningTime="2025-11-23 20:27:25.135039121 +0000 UTC m=+1153.284080077" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.158978 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.262090 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw4cp\" (UniqueName: \"kubernetes.io/projected/847e98eb-74a9-4936-990a-881fbc0b2271-kube-api-access-qw4cp\") pod \"847e98eb-74a9-4936-990a-881fbc0b2271\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.262175 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-config-data\") pod \"847e98eb-74a9-4936-990a-881fbc0b2271\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.262345 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-combined-ca-bundle\") pod \"847e98eb-74a9-4936-990a-881fbc0b2271\" (UID: \"847e98eb-74a9-4936-990a-881fbc0b2271\") " Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.272069 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/847e98eb-74a9-4936-990a-881fbc0b2271-kube-api-access-qw4cp" (OuterVolumeSpecName: "kube-api-access-qw4cp") pod "847e98eb-74a9-4936-990a-881fbc0b2271" (UID: "847e98eb-74a9-4936-990a-881fbc0b2271"). InnerVolumeSpecName "kube-api-access-qw4cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.302764 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-config-data" (OuterVolumeSpecName: "config-data") pod "847e98eb-74a9-4936-990a-881fbc0b2271" (UID: "847e98eb-74a9-4936-990a-881fbc0b2271"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.340104 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "847e98eb-74a9-4936-990a-881fbc0b2271" (UID: "847e98eb-74a9-4936-990a-881fbc0b2271"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.365174 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw4cp\" (UniqueName: \"kubernetes.io/projected/847e98eb-74a9-4936-990a-881fbc0b2271-kube-api-access-qw4cp\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.365217 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:25 crc kubenswrapper[4726]: I1123 20:27:25.365227 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/847e98eb-74a9-4936-990a-881fbc0b2271-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.120292 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.150724 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.162328 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.185663 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:26 crc kubenswrapper[4726]: E1123 20:27:26.186372 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="847e98eb-74a9-4936-990a-881fbc0b2271" containerName="nova-scheduler-scheduler" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.186443 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="847e98eb-74a9-4936-990a-881fbc0b2271" containerName="nova-scheduler-scheduler" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.187013 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="847e98eb-74a9-4936-990a-881fbc0b2271" containerName="nova-scheduler-scheduler" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.187965 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.190740 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.195533 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwvhc\" (UniqueName: \"kubernetes.io/projected/a310fd7f-5f37-4428-ba58-a8f4184ce283-kube-api-access-lwvhc\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.195595 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.195683 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-config-data\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.232596 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.297088 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwvhc\" (UniqueName: \"kubernetes.io/projected/a310fd7f-5f37-4428-ba58-a8f4184ce283-kube-api-access-lwvhc\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.297144 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.297200 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-config-data\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.302528 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-config-data\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.303037 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.320450 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwvhc\" (UniqueName: \"kubernetes.io/projected/a310fd7f-5f37-4428-ba58-a8f4184ce283-kube-api-access-lwvhc\") pod \"nova-scheduler-0\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.525374 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:27:26 crc kubenswrapper[4726]: I1123 20:27:26.603834 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="847e98eb-74a9-4936-990a-881fbc0b2271" path="/var/lib/kubelet/pods/847e98eb-74a9-4936-990a-881fbc0b2271/volumes" Nov 23 20:27:27 crc kubenswrapper[4726]: W1123 20:27:27.123990 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda310fd7f_5f37_4428_ba58_a8f4184ce283.slice/crio-f9f35cb839998f0bc61dd01ad027746eca9efb1922a97a280d054e04a5998667 WatchSource:0}: Error finding container f9f35cb839998f0bc61dd01ad027746eca9efb1922a97a280d054e04a5998667: Status 404 returned error can't find the container with id f9f35cb839998f0bc61dd01ad027746eca9efb1922a97a280d054e04a5998667 Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.124630 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.143261 4726 generic.go:334] "Generic (PLEG): container finished" podID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerID="4284d64b096fcdd9d651c217914f79124327dd0d66f13378e0c7718153fae473" exitCode=0 Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.143353 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99b668ba-4608-4fa3-8342-98f1af7f5263","Type":"ContainerDied","Data":"4284d64b096fcdd9d651c217914f79124327dd0d66f13378e0c7718153fae473"} Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.265599 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.327616 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-config-data\") pod \"99b668ba-4608-4fa3-8342-98f1af7f5263\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.327709 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b668ba-4608-4fa3-8342-98f1af7f5263-logs\") pod \"99b668ba-4608-4fa3-8342-98f1af7f5263\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.327765 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdwt8\" (UniqueName: \"kubernetes.io/projected/99b668ba-4608-4fa3-8342-98f1af7f5263-kube-api-access-zdwt8\") pod \"99b668ba-4608-4fa3-8342-98f1af7f5263\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.327816 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-combined-ca-bundle\") pod \"99b668ba-4608-4fa3-8342-98f1af7f5263\" (UID: \"99b668ba-4608-4fa3-8342-98f1af7f5263\") " Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.329162 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99b668ba-4608-4fa3-8342-98f1af7f5263-logs" (OuterVolumeSpecName: "logs") pod "99b668ba-4608-4fa3-8342-98f1af7f5263" (UID: "99b668ba-4608-4fa3-8342-98f1af7f5263"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.346122 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b668ba-4608-4fa3-8342-98f1af7f5263-kube-api-access-zdwt8" (OuterVolumeSpecName: "kube-api-access-zdwt8") pod "99b668ba-4608-4fa3-8342-98f1af7f5263" (UID: "99b668ba-4608-4fa3-8342-98f1af7f5263"). InnerVolumeSpecName "kube-api-access-zdwt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.367052 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-config-data" (OuterVolumeSpecName: "config-data") pod "99b668ba-4608-4fa3-8342-98f1af7f5263" (UID: "99b668ba-4608-4fa3-8342-98f1af7f5263"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.375116 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99b668ba-4608-4fa3-8342-98f1af7f5263" (UID: "99b668ba-4608-4fa3-8342-98f1af7f5263"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.429714 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdwt8\" (UniqueName: \"kubernetes.io/projected/99b668ba-4608-4fa3-8342-98f1af7f5263-kube-api-access-zdwt8\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.429765 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.429774 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b668ba-4608-4fa3-8342-98f1af7f5263-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:27 crc kubenswrapper[4726]: I1123 20:27:27.429782 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b668ba-4608-4fa3-8342-98f1af7f5263-logs\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.152909 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a310fd7f-5f37-4428-ba58-a8f4184ce283","Type":"ContainerStarted","Data":"26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152"} Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.154280 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a310fd7f-5f37-4428-ba58-a8f4184ce283","Type":"ContainerStarted","Data":"f9f35cb839998f0bc61dd01ad027746eca9efb1922a97a280d054e04a5998667"} Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.154996 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.154985 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"99b668ba-4608-4fa3-8342-98f1af7f5263","Type":"ContainerDied","Data":"6edda3020a88e01aca489c2d2d8cab7d64a0170264e3a2cf854d19ac3bd9cfb7"} Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.155211 4726 scope.go:117] "RemoveContainer" containerID="4284d64b096fcdd9d651c217914f79124327dd0d66f13378e0c7718153fae473" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.192230 4726 scope.go:117] "RemoveContainer" containerID="8da6a9027874c9ce99e53b9df7d1f66f74dd84c747f9445e3d8b63049c9e7513" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.203271 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.203244193 podStartE2EDuration="2.203244193s" podCreationTimestamp="2025-11-23 20:27:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:28.175907824 +0000 UTC m=+1156.324948790" watchObservedRunningTime="2025-11-23 20:27:28.203244193 +0000 UTC m=+1156.352285149" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.216421 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.245584 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.277988 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:28 crc kubenswrapper[4726]: E1123 20:27:28.279036 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-log" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.279057 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-log" Nov 23 20:27:28 crc kubenswrapper[4726]: E1123 20:27:28.279082 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-api" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.279091 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-api" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.280149 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-api" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.280178 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" containerName="nova-api-log" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.282385 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.296009 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.323895 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.347761 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-config-data\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.347816 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62224881-e65a-4572-b8ce-f5f538411754-logs\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.347898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52xbk\" (UniqueName: \"kubernetes.io/projected/62224881-e65a-4572-b8ce-f5f538411754-kube-api-access-52xbk\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.347928 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.449429 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52xbk\" (UniqueName: \"kubernetes.io/projected/62224881-e65a-4572-b8ce-f5f538411754-kube-api-access-52xbk\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.449483 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.449566 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-config-data\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.449597 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62224881-e65a-4572-b8ce-f5f538411754-logs\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.449981 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62224881-e65a-4572-b8ce-f5f538411754-logs\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.454738 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.455429 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-config-data\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.467248 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52xbk\" (UniqueName: \"kubernetes.io/projected/62224881-e65a-4572-b8ce-f5f538411754-kube-api-access-52xbk\") pod \"nova-api-0\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " pod="openstack/nova-api-0" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.600599 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99b668ba-4608-4fa3-8342-98f1af7f5263" path="/var/lib/kubelet/pods/99b668ba-4608-4fa3-8342-98f1af7f5263/volumes" Nov 23 20:27:28 crc kubenswrapper[4726]: I1123 20:27:28.632174 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:29 crc kubenswrapper[4726]: I1123 20:27:29.097591 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:29 crc kubenswrapper[4726]: W1123 20:27:29.103167 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62224881_e65a_4572_b8ce_f5f538411754.slice/crio-758cfe9f9ad125a101ef7c171a43faa5d9841f14e6b5b84fc3b2282fb4d98ba7 WatchSource:0}: Error finding container 758cfe9f9ad125a101ef7c171a43faa5d9841f14e6b5b84fc3b2282fb4d98ba7: Status 404 returned error can't find the container with id 758cfe9f9ad125a101ef7c171a43faa5d9841f14e6b5b84fc3b2282fb4d98ba7 Nov 23 20:27:29 crc kubenswrapper[4726]: I1123 20:27:29.173804 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62224881-e65a-4572-b8ce-f5f538411754","Type":"ContainerStarted","Data":"758cfe9f9ad125a101ef7c171a43faa5d9841f14e6b5b84fc3b2282fb4d98ba7"} Nov 23 20:27:30 crc kubenswrapper[4726]: I1123 20:27:30.182631 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62224881-e65a-4572-b8ce-f5f538411754","Type":"ContainerStarted","Data":"770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70"} Nov 23 20:27:30 crc kubenswrapper[4726]: I1123 20:27:30.182923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62224881-e65a-4572-b8ce-f5f538411754","Type":"ContainerStarted","Data":"f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212"} Nov 23 20:27:30 crc kubenswrapper[4726]: I1123 20:27:30.214666 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.214647383 podStartE2EDuration="2.214647383s" podCreationTimestamp="2025-11-23 20:27:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:30.20930187 +0000 UTC m=+1158.358342846" watchObservedRunningTime="2025-11-23 20:27:30.214647383 +0000 UTC m=+1158.363688339" Nov 23 20:27:31 crc kubenswrapper[4726]: I1123 20:27:31.527058 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 20:27:33 crc kubenswrapper[4726]: I1123 20:27:33.499395 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 23 20:27:36 crc kubenswrapper[4726]: I1123 20:27:36.526455 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 20:27:36 crc kubenswrapper[4726]: I1123 20:27:36.561008 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 20:27:37 crc kubenswrapper[4726]: I1123 20:27:37.320354 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 23 20:27:38 crc kubenswrapper[4726]: I1123 20:27:38.633710 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 20:27:38 crc kubenswrapper[4726]: I1123 20:27:38.634008 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 20:27:39 crc kubenswrapper[4726]: I1123 20:27:39.716078 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.179:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 20:27:39 crc kubenswrapper[4726]: I1123 20:27:39.716803 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.179:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.114544 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.123675 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.234436 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrrpv\" (UniqueName: \"kubernetes.io/projected/d771b2fa-fa53-48f5-ae54-112d535e04a3-kube-api-access-lrrpv\") pod \"d771b2fa-fa53-48f5-ae54-112d535e04a3\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.234854 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcrs8\" (UniqueName: \"kubernetes.io/projected/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-kube-api-access-qcrs8\") pod \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.235019 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-config-data\") pod \"d771b2fa-fa53-48f5-ae54-112d535e04a3\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.235187 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-logs\") pod \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.235297 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-combined-ca-bundle\") pod \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.235434 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-combined-ca-bundle\") pod \"d771b2fa-fa53-48f5-ae54-112d535e04a3\" (UID: \"d771b2fa-fa53-48f5-ae54-112d535e04a3\") " Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.235503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-logs" (OuterVolumeSpecName: "logs") pod "5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" (UID: "5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.235509 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-config-data\") pod \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\" (UID: \"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8\") " Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.240727 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d771b2fa-fa53-48f5-ae54-112d535e04a3-kube-api-access-lrrpv" (OuterVolumeSpecName: "kube-api-access-lrrpv") pod "d771b2fa-fa53-48f5-ae54-112d535e04a3" (UID: "d771b2fa-fa53-48f5-ae54-112d535e04a3"). InnerVolumeSpecName "kube-api-access-lrrpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.242158 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrrpv\" (UniqueName: \"kubernetes.io/projected/d771b2fa-fa53-48f5-ae54-112d535e04a3-kube-api-access-lrrpv\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.242242 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-logs\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.263543 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-config-data" (OuterVolumeSpecName: "config-data") pod "d771b2fa-fa53-48f5-ae54-112d535e04a3" (UID: "d771b2fa-fa53-48f5-ae54-112d535e04a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.268567 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-kube-api-access-qcrs8" (OuterVolumeSpecName: "kube-api-access-qcrs8") pod "5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" (UID: "5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8"). InnerVolumeSpecName "kube-api-access-qcrs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.275846 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" (UID: "5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.286804 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-config-data" (OuterVolumeSpecName: "config-data") pod "5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" (UID: "5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.293516 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d771b2fa-fa53-48f5-ae54-112d535e04a3" (UID: "d771b2fa-fa53-48f5-ae54-112d535e04a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.344035 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.344074 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.344089 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d771b2fa-fa53-48f5-ae54-112d535e04a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.344106 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.344119 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcrs8\" (UniqueName: \"kubernetes.io/projected/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8-kube-api-access-qcrs8\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.350404 4726 generic.go:334] "Generic (PLEG): container finished" podID="d771b2fa-fa53-48f5-ae54-112d535e04a3" containerID="2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f" exitCode=137 Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.350579 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.350959 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d771b2fa-fa53-48f5-ae54-112d535e04a3","Type":"ContainerDied","Data":"2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f"} Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.351004 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d771b2fa-fa53-48f5-ae54-112d535e04a3","Type":"ContainerDied","Data":"21b5277fe418174dbae3716a0c141ae33d3e360733c0e7d055826649d409aa73"} Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.351043 4726 scope.go:117] "RemoveContainer" containerID="2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.353448 4726 generic.go:334] "Generic (PLEG): container finished" podID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerID="e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753" exitCode=137 Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.353476 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.353498 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8","Type":"ContainerDied","Data":"e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753"} Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.353528 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8","Type":"ContainerDied","Data":"ff72b0e8725bef3c65f26e81ad937c44537113d6211535408f9500d933c25450"} Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.376463 4726 scope.go:117] "RemoveContainer" containerID="2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f" Nov 23 20:27:43 crc kubenswrapper[4726]: E1123 20:27:43.377351 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f\": container with ID starting with 2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f not found: ID does not exist" containerID="2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.377391 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f"} err="failed to get container status \"2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f\": rpc error: code = NotFound desc = could not find container \"2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f\": container with ID starting with 2389c4e952bec853c25222dccbadcc2389898a753acc5371880e6794c573d24f not found: ID does not exist" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.377452 4726 scope.go:117] "RemoveContainer" containerID="e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.406975 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.424880 4726 scope.go:117] "RemoveContainer" containerID="1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.434474 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.461177 4726 scope.go:117] "RemoveContainer" containerID="e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753" Nov 23 20:27:43 crc kubenswrapper[4726]: E1123 20:27:43.464410 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753\": container with ID starting with e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753 not found: ID does not exist" containerID="e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.464539 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753"} err="failed to get container status \"e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753\": rpc error: code = NotFound desc = could not find container \"e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753\": container with ID starting with e29a225c77070dd04e734d986a47b12cfcfc705cdb8d55cf1c8efb8043d2a753 not found: ID does not exist" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.464692 4726 scope.go:117] "RemoveContainer" containerID="1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491" Nov 23 20:27:43 crc kubenswrapper[4726]: E1123 20:27:43.467036 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491\": container with ID starting with 1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491 not found: ID does not exist" containerID="1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.467179 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491"} err="failed to get container status \"1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491\": rpc error: code = NotFound desc = could not find container \"1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491\": container with ID starting with 1c690b812b88bc6b58ca6eec6c93ec0aa63a14de1ec7a8e2ad5a1f5d2f9d7491 not found: ID does not exist" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.468467 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.495067 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.504125 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:43 crc kubenswrapper[4726]: E1123 20:27:43.504564 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d771b2fa-fa53-48f5-ae54-112d535e04a3" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.504581 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d771b2fa-fa53-48f5-ae54-112d535e04a3" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 20:27:43 crc kubenswrapper[4726]: E1123 20:27:43.504591 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerName="nova-metadata-log" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.504598 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerName="nova-metadata-log" Nov 23 20:27:43 crc kubenswrapper[4726]: E1123 20:27:43.504610 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerName="nova-metadata-metadata" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.504617 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerName="nova-metadata-metadata" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.504792 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d771b2fa-fa53-48f5-ae54-112d535e04a3" containerName="nova-cell1-novncproxy-novncproxy" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.504814 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerName="nova-metadata-log" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.504828 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" containerName="nova-metadata-metadata" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.505483 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.508615 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.508721 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.508725 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.515483 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.528726 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.530504 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.532753 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.536717 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.537316 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.551248 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.551308 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwj7c\" (UniqueName: \"kubernetes.io/projected/364f0735-33a2-4a68-88ac-86842a3a5738-kube-api-access-zwj7c\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.552085 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.552138 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p26dq\" (UniqueName: \"kubernetes.io/projected/609ae5f6-90fc-432a-820a-ca996f78c1c7-kube-api-access-p26dq\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.552159 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-config-data\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.552175 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.552198 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.552228 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.552256 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364f0735-33a2-4a68-88ac-86842a3a5738-logs\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.552471 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654101 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364f0735-33a2-4a68-88ac-86842a3a5738-logs\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654521 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654537 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364f0735-33a2-4a68-88ac-86842a3a5738-logs\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654582 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654634 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwj7c\" (UniqueName: \"kubernetes.io/projected/364f0735-33a2-4a68-88ac-86842a3a5738-kube-api-access-zwj7c\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654701 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654763 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p26dq\" (UniqueName: \"kubernetes.io/projected/609ae5f6-90fc-432a-820a-ca996f78c1c7-kube-api-access-p26dq\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654793 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-config-data\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654810 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654835 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.654938 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.659161 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.660030 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.661593 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.661604 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.661743 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-config-data\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.662417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.662790 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/609ae5f6-90fc-432a-820a-ca996f78c1c7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.670916 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p26dq\" (UniqueName: \"kubernetes.io/projected/609ae5f6-90fc-432a-820a-ca996f78c1c7-kube-api-access-p26dq\") pod \"nova-cell1-novncproxy-0\" (UID: \"609ae5f6-90fc-432a-820a-ca996f78c1c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.671398 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwj7c\" (UniqueName: \"kubernetes.io/projected/364f0735-33a2-4a68-88ac-86842a3a5738-kube-api-access-zwj7c\") pod \"nova-metadata-0\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " pod="openstack/nova-metadata-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.823306 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:43 crc kubenswrapper[4726]: I1123 20:27:43.848217 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:27:44 crc kubenswrapper[4726]: I1123 20:27:44.334305 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 23 20:27:44 crc kubenswrapper[4726]: I1123 20:27:44.366434 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"609ae5f6-90fc-432a-820a-ca996f78c1c7","Type":"ContainerStarted","Data":"93a8533ee4f760b6f0aaab909663509d6e8a2b68baddae385ad72a7b271e5b17"} Nov 23 20:27:44 crc kubenswrapper[4726]: I1123 20:27:44.394894 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:27:44 crc kubenswrapper[4726]: I1123 20:27:44.600653 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8" path="/var/lib/kubelet/pods/5ae2ce8d-ca20-4cd2-810b-a5b24ab623c8/volumes" Nov 23 20:27:44 crc kubenswrapper[4726]: I1123 20:27:44.601570 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d771b2fa-fa53-48f5-ae54-112d535e04a3" path="/var/lib/kubelet/pods/d771b2fa-fa53-48f5-ae54-112d535e04a3/volumes" Nov 23 20:27:45 crc kubenswrapper[4726]: I1123 20:27:45.387578 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"609ae5f6-90fc-432a-820a-ca996f78c1c7","Type":"ContainerStarted","Data":"301c12fb583152b5a22dcff31f73127f4ca050c948e7163d1c34c33b37041fa7"} Nov 23 20:27:45 crc kubenswrapper[4726]: I1123 20:27:45.394971 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"364f0735-33a2-4a68-88ac-86842a3a5738","Type":"ContainerStarted","Data":"ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a"} Nov 23 20:27:45 crc kubenswrapper[4726]: I1123 20:27:45.395013 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"364f0735-33a2-4a68-88ac-86842a3a5738","Type":"ContainerStarted","Data":"9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe"} Nov 23 20:27:45 crc kubenswrapper[4726]: I1123 20:27:45.395024 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"364f0735-33a2-4a68-88ac-86842a3a5738","Type":"ContainerStarted","Data":"bdf1c953be13c061d6221e533a8c02887a55cc982c00228598d52f4c517192dc"} Nov 23 20:27:45 crc kubenswrapper[4726]: I1123 20:27:45.414091 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.414004004 podStartE2EDuration="2.414004004s" podCreationTimestamp="2025-11-23 20:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:45.409553291 +0000 UTC m=+1173.558594257" watchObservedRunningTime="2025-11-23 20:27:45.414004004 +0000 UTC m=+1173.563044960" Nov 23 20:27:45 crc kubenswrapper[4726]: I1123 20:27:45.431221 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.43119801 podStartE2EDuration="2.43119801s" podCreationTimestamp="2025-11-23 20:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:45.429266797 +0000 UTC m=+1173.578307793" watchObservedRunningTime="2025-11-23 20:27:45.43119801 +0000 UTC m=+1173.580238966" Nov 23 20:27:48 crc kubenswrapper[4726]: I1123 20:27:48.637516 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 20:27:48 crc kubenswrapper[4726]: I1123 20:27:48.639226 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 20:27:48 crc kubenswrapper[4726]: I1123 20:27:48.639700 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 20:27:48 crc kubenswrapper[4726]: I1123 20:27:48.642054 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 20:27:48 crc kubenswrapper[4726]: I1123 20:27:48.824717 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:48 crc kubenswrapper[4726]: I1123 20:27:48.848390 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 20:27:48 crc kubenswrapper[4726]: I1123 20:27:48.848460 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.353947 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.437751 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.441460 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.645780 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c9b558957-bzgd2"] Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.658552 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.685551 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-bzgd2"] Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.811404 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-config\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.811529 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqnmj\" (UniqueName: \"kubernetes.io/projected/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-kube-api-access-mqnmj\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.811592 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.811626 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-dns-svc\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.811674 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.913056 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.913145 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-dns-svc\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.913171 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.913243 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-config\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.913283 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqnmj\" (UniqueName: \"kubernetes.io/projected/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-kube-api-access-mqnmj\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.914390 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-nb\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.914906 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-dns-svc\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.915408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-sb\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.915910 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-config\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:49 crc kubenswrapper[4726]: I1123 20:27:49.935947 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqnmj\" (UniqueName: \"kubernetes.io/projected/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-kube-api-access-mqnmj\") pod \"dnsmasq-dns-c9b558957-bzgd2\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:50 crc kubenswrapper[4726]: I1123 20:27:50.029432 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:50 crc kubenswrapper[4726]: I1123 20:27:50.503083 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-bzgd2"] Nov 23 20:27:51 crc kubenswrapper[4726]: I1123 20:27:51.471181 4726 generic.go:334] "Generic (PLEG): container finished" podID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" containerID="efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9" exitCode=0 Nov 23 20:27:51 crc kubenswrapper[4726]: I1123 20:27:51.471442 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" event={"ID":"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a","Type":"ContainerDied","Data":"efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9"} Nov 23 20:27:51 crc kubenswrapper[4726]: I1123 20:27:51.472028 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" event={"ID":"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a","Type":"ContainerStarted","Data":"d73f428227a5f562f2b18cecc50b6363753595e1a1707fcb13c742acfe5ab37c"} Nov 23 20:27:51 crc kubenswrapper[4726]: I1123 20:27:51.987814 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.214892 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.215261 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="ceilometer-central-agent" containerID="cri-o://d09d262f5944c3ae1d4472ee5dc1c1861c28e6609a121309e5e3cdf77b9026b8" gracePeriod=30 Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.215323 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="proxy-httpd" containerID="cri-o://c32e4b5a321337406904d31573f68e83c5cb223948120df2db90f9738e3c7048" gracePeriod=30 Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.215356 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="sg-core" containerID="cri-o://8c21b267f39bb05298b4ae30e5df047d335ab6c459be6d35162237db33f72bc4" gracePeriod=30 Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.215583 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="ceilometer-notification-agent" containerID="cri-o://b588c404918fc83ab4604909fba7fef861124df74fbfa6a0db6c600dd4f65366" gracePeriod=30 Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.485080 4726 generic.go:334] "Generic (PLEG): container finished" podID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerID="8c21b267f39bb05298b4ae30e5df047d335ab6c459be6d35162237db33f72bc4" exitCode=2 Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.485153 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerDied","Data":"8c21b267f39bb05298b4ae30e5df047d335ab6c459be6d35162237db33f72bc4"} Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.488354 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" event={"ID":"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a","Type":"ContainerStarted","Data":"b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771"} Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.488545 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.488522 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-log" containerID="cri-o://f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212" gracePeriod=30 Nov 23 20:27:52 crc kubenswrapper[4726]: I1123 20:27:52.488672 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-api" containerID="cri-o://770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70" gracePeriod=30 Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.505809 4726 generic.go:334] "Generic (PLEG): container finished" podID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerID="c32e4b5a321337406904d31573f68e83c5cb223948120df2db90f9738e3c7048" exitCode=0 Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.506310 4726 generic.go:334] "Generic (PLEG): container finished" podID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerID="d09d262f5944c3ae1d4472ee5dc1c1861c28e6609a121309e5e3cdf77b9026b8" exitCode=0 Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.505994 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerDied","Data":"c32e4b5a321337406904d31573f68e83c5cb223948120df2db90f9738e3c7048"} Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.506400 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerDied","Data":"d09d262f5944c3ae1d4472ee5dc1c1861c28e6609a121309e5e3cdf77b9026b8"} Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.511355 4726 generic.go:334] "Generic (PLEG): container finished" podID="62224881-e65a-4572-b8ce-f5f538411754" containerID="f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212" exitCode=143 Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.512007 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62224881-e65a-4572-b8ce-f5f538411754","Type":"ContainerDied","Data":"f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212"} Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.823596 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.848486 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.848537 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.851922 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:53 crc kubenswrapper[4726]: I1123 20:27:53.885130 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" podStartSLOduration=4.885111461 podStartE2EDuration="4.885111461s" podCreationTimestamp="2025-11-23 20:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:52.518983592 +0000 UTC m=+1180.668024558" watchObservedRunningTime="2025-11-23 20:27:53.885111461 +0000 UTC m=+1182.034152417" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.541248 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.728613 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-8qvxw"] Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.729960 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.735303 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.738927 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8qvxw"] Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.739255 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.845501 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.845601 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmjp6\" (UniqueName: \"kubernetes.io/projected/267f6c6b-3b5d-4e89-bf80-3650b890aed9-kube-api-access-jmjp6\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.845683 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-scripts\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.845705 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-config-data\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.864068 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.864089 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.947357 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmjp6\" (UniqueName: \"kubernetes.io/projected/267f6c6b-3b5d-4e89-bf80-3650b890aed9-kube-api-access-jmjp6\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.947485 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-scripts\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.947512 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-config-data\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.947564 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.956267 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-config-data\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.962583 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.963001 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-scripts\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:54 crc kubenswrapper[4726]: I1123 20:27:54.964474 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmjp6\" (UniqueName: \"kubernetes.io/projected/267f6c6b-3b5d-4e89-bf80-3650b890aed9-kube-api-access-jmjp6\") pod \"nova-cell1-cell-mapping-8qvxw\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.071442 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.539370 4726 generic.go:334] "Generic (PLEG): container finished" podID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerID="b588c404918fc83ab4604909fba7fef861124df74fbfa6a0db6c600dd4f65366" exitCode=0 Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.540891 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerDied","Data":"b588c404918fc83ab4604909fba7fef861124df74fbfa6a0db6c600dd4f65366"} Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.690266 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8qvxw"] Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.776015 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.885500 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-ceilometer-tls-certs\") pod \"c4cfd52a-db3c-4f20-89b3-a29921534365\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.885576 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-sg-core-conf-yaml\") pod \"c4cfd52a-db3c-4f20-89b3-a29921534365\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.885610 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-log-httpd\") pod \"c4cfd52a-db3c-4f20-89b3-a29921534365\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.885682 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-config-data\") pod \"c4cfd52a-db3c-4f20-89b3-a29921534365\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.885750 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-run-httpd\") pod \"c4cfd52a-db3c-4f20-89b3-a29921534365\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.885781 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-combined-ca-bundle\") pod \"c4cfd52a-db3c-4f20-89b3-a29921534365\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.885826 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-scripts\") pod \"c4cfd52a-db3c-4f20-89b3-a29921534365\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.885856 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rmw7\" (UniqueName: \"kubernetes.io/projected/c4cfd52a-db3c-4f20-89b3-a29921534365-kube-api-access-8rmw7\") pod \"c4cfd52a-db3c-4f20-89b3-a29921534365\" (UID: \"c4cfd52a-db3c-4f20-89b3-a29921534365\") " Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.888238 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c4cfd52a-db3c-4f20-89b3-a29921534365" (UID: "c4cfd52a-db3c-4f20-89b3-a29921534365"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.888539 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c4cfd52a-db3c-4f20-89b3-a29921534365" (UID: "c4cfd52a-db3c-4f20-89b3-a29921534365"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.892210 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-scripts" (OuterVolumeSpecName: "scripts") pod "c4cfd52a-db3c-4f20-89b3-a29921534365" (UID: "c4cfd52a-db3c-4f20-89b3-a29921534365"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.894222 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4cfd52a-db3c-4f20-89b3-a29921534365-kube-api-access-8rmw7" (OuterVolumeSpecName: "kube-api-access-8rmw7") pod "c4cfd52a-db3c-4f20-89b3-a29921534365" (UID: "c4cfd52a-db3c-4f20-89b3-a29921534365"). InnerVolumeSpecName "kube-api-access-8rmw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.988658 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.988682 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.988693 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rmw7\" (UniqueName: \"kubernetes.io/projected/c4cfd52a-db3c-4f20-89b3-a29921534365-kube-api-access-8rmw7\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:55 crc kubenswrapper[4726]: I1123 20:27:55.988702 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4cfd52a-db3c-4f20-89b3-a29921534365-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.002091 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c4cfd52a-db3c-4f20-89b3-a29921534365" (UID: "c4cfd52a-db3c-4f20-89b3-a29921534365"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.006823 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c4cfd52a-db3c-4f20-89b3-a29921534365" (UID: "c4cfd52a-db3c-4f20-89b3-a29921534365"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.016195 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4cfd52a-db3c-4f20-89b3-a29921534365" (UID: "c4cfd52a-db3c-4f20-89b3-a29921534365"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.070044 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-config-data" (OuterVolumeSpecName: "config-data") pod "c4cfd52a-db3c-4f20-89b3-a29921534365" (UID: "c4cfd52a-db3c-4f20-89b3-a29921534365"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.091464 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.091498 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.091509 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.091517 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4cfd52a-db3c-4f20-89b3-a29921534365-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.407357 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.498807 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52xbk\" (UniqueName: \"kubernetes.io/projected/62224881-e65a-4572-b8ce-f5f538411754-kube-api-access-52xbk\") pod \"62224881-e65a-4572-b8ce-f5f538411754\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.498879 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62224881-e65a-4572-b8ce-f5f538411754-logs\") pod \"62224881-e65a-4572-b8ce-f5f538411754\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.498963 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-combined-ca-bundle\") pod \"62224881-e65a-4572-b8ce-f5f538411754\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.499034 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-config-data\") pod \"62224881-e65a-4572-b8ce-f5f538411754\" (UID: \"62224881-e65a-4572-b8ce-f5f538411754\") " Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.501319 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62224881-e65a-4572-b8ce-f5f538411754-logs" (OuterVolumeSpecName: "logs") pod "62224881-e65a-4572-b8ce-f5f538411754" (UID: "62224881-e65a-4572-b8ce-f5f538411754"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.509662 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62224881-e65a-4572-b8ce-f5f538411754-kube-api-access-52xbk" (OuterVolumeSpecName: "kube-api-access-52xbk") pod "62224881-e65a-4572-b8ce-f5f538411754" (UID: "62224881-e65a-4572-b8ce-f5f538411754"). InnerVolumeSpecName "kube-api-access-52xbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.568479 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62224881-e65a-4572-b8ce-f5f538411754" (UID: "62224881-e65a-4572-b8ce-f5f538411754"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.577126 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-config-data" (OuterVolumeSpecName: "config-data") pod "62224881-e65a-4572-b8ce-f5f538411754" (UID: "62224881-e65a-4572-b8ce-f5f538411754"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.582546 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4cfd52a-db3c-4f20-89b3-a29921534365","Type":"ContainerDied","Data":"ed290acc71239d479e7b960931260b18be2b04fdaa289a71bb992dac708cecf2"} Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.582605 4726 scope.go:117] "RemoveContainer" containerID="c32e4b5a321337406904d31573f68e83c5cb223948120df2db90f9738e3c7048" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.582768 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.603355 4726 generic.go:334] "Generic (PLEG): container finished" podID="62224881-e65a-4572-b8ce-f5f538411754" containerID="770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70" exitCode=0 Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.603517 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.606986 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52xbk\" (UniqueName: \"kubernetes.io/projected/62224881-e65a-4572-b8ce-f5f538411754-kube-api-access-52xbk\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.607016 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62224881-e65a-4572-b8ce-f5f538411754-logs\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.607029 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.607039 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62224881-e65a-4572-b8ce-f5f538411754-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.612213 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8qvxw" event={"ID":"267f6c6b-3b5d-4e89-bf80-3650b890aed9","Type":"ContainerStarted","Data":"cff0b94346727804ad9dcbf63d396db0137c49417b03f512ba5091ab5bb59225"} Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.612260 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8qvxw" event={"ID":"267f6c6b-3b5d-4e89-bf80-3650b890aed9","Type":"ContainerStarted","Data":"e57ff1d16aaab28cf695f973cda52b180f43fc46e60b1c34b69e33dd449891fc"} Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.612275 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62224881-e65a-4572-b8ce-f5f538411754","Type":"ContainerDied","Data":"770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70"} Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.612293 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62224881-e65a-4572-b8ce-f5f538411754","Type":"ContainerDied","Data":"758cfe9f9ad125a101ef7c171a43faa5d9841f14e6b5b84fc3b2282fb4d98ba7"} Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.640361 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-8qvxw" podStartSLOduration=2.640334215 podStartE2EDuration="2.640334215s" podCreationTimestamp="2025-11-23 20:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:56.628974371 +0000 UTC m=+1184.778015327" watchObservedRunningTime="2025-11-23 20:27:56.640334215 +0000 UTC m=+1184.789375171" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.652577 4726 scope.go:117] "RemoveContainer" containerID="8c21b267f39bb05298b4ae30e5df047d335ab6c459be6d35162237db33f72bc4" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.688946 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.703959 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.719741 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:56 crc kubenswrapper[4726]: E1123 20:27:56.720215 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-log" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720230 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-log" Nov 23 20:27:56 crc kubenswrapper[4726]: E1123 20:27:56.720244 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="ceilometer-notification-agent" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720253 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="ceilometer-notification-agent" Nov 23 20:27:56 crc kubenswrapper[4726]: E1123 20:27:56.720289 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="proxy-httpd" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720298 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="proxy-httpd" Nov 23 20:27:56 crc kubenswrapper[4726]: E1123 20:27:56.720311 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="ceilometer-central-agent" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720317 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="ceilometer-central-agent" Nov 23 20:27:56 crc kubenswrapper[4726]: E1123 20:27:56.720333 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="sg-core" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720339 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="sg-core" Nov 23 20:27:56 crc kubenswrapper[4726]: E1123 20:27:56.720357 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-api" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720365 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-api" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720639 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-api" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720663 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="ceilometer-notification-agent" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720678 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="ceilometer-central-agent" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720686 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="sg-core" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720697 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" containerName="proxy-httpd" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.720713 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="62224881-e65a-4572-b8ce-f5f538411754" containerName="nova-api-log" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.725842 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.730114 4726 scope.go:117] "RemoveContainer" containerID="b588c404918fc83ab4604909fba7fef861124df74fbfa6a0db6c600dd4f65366" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.730216 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.732824 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.733120 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.733822 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.753320 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.765191 4726 scope.go:117] "RemoveContainer" containerID="d09d262f5944c3ae1d4472ee5dc1c1861c28e6609a121309e5e3cdf77b9026b8" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.766935 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.775045 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.778006 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.782030 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.782184 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.782267 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.784750 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.797712 4726 scope.go:117] "RemoveContainer" containerID="770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.813858 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-run-httpd\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.814000 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-config-data\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.814049 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.814090 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-log-httpd\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.814136 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q27v5\" (UniqueName: \"kubernetes.io/projected/6b15cde8-6249-48c9-a7f3-984adb91acf1-kube-api-access-q27v5\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.814205 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-scripts\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.814234 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.814276 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.825203 4726 scope.go:117] "RemoveContainer" containerID="f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.863788 4726 scope.go:117] "RemoveContainer" containerID="770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70" Nov 23 20:27:56 crc kubenswrapper[4726]: E1123 20:27:56.864609 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70\": container with ID starting with 770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70 not found: ID does not exist" containerID="770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.864643 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70"} err="failed to get container status \"770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70\": rpc error: code = NotFound desc = could not find container \"770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70\": container with ID starting with 770a1f861b4b3d805c76d3336b0d4d0f20bd75d02e29b125edc9a889de80cc70 not found: ID does not exist" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.864666 4726 scope.go:117] "RemoveContainer" containerID="f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212" Nov 23 20:27:56 crc kubenswrapper[4726]: E1123 20:27:56.865129 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212\": container with ID starting with f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212 not found: ID does not exist" containerID="f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.865192 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212"} err="failed to get container status \"f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212\": rpc error: code = NotFound desc = could not find container \"f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212\": container with ID starting with f52beeeecf83f83106015bb1118096895537858a3289e9ed70523db5a7703212 not found: ID does not exist" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916448 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916568 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916599 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-log-httpd\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916641 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q27v5\" (UniqueName: \"kubernetes.io/projected/6b15cde8-6249-48c9-a7f3-984adb91acf1-kube-api-access-q27v5\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916709 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2lqf\" (UniqueName: \"kubernetes.io/projected/218ec494-3f53-4cb7-8d33-10f7824f4e7c-kube-api-access-x2lqf\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916755 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-scripts\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-config-data\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916801 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-public-tls-certs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916832 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916881 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916934 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-run-httpd\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.916971 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218ec494-3f53-4cb7-8d33-10f7824f4e7c-logs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.917019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.917046 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-config-data\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.918792 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-log-httpd\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.918978 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-run-httpd\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.921659 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.922021 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-scripts\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.923006 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.923223 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-config-data\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.932582 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:56 crc kubenswrapper[4726]: I1123 20:27:56.934998 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q27v5\" (UniqueName: \"kubernetes.io/projected/6b15cde8-6249-48c9-a7f3-984adb91acf1-kube-api-access-q27v5\") pod \"ceilometer-0\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " pod="openstack/ceilometer-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.018712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.018800 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.018888 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2lqf\" (UniqueName: \"kubernetes.io/projected/218ec494-3f53-4cb7-8d33-10f7824f4e7c-kube-api-access-x2lqf\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.018915 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-config-data\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.018934 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-public-tls-certs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.019719 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218ec494-3f53-4cb7-8d33-10f7824f4e7c-logs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.020180 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218ec494-3f53-4cb7-8d33-10f7824f4e7c-logs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.022530 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.023250 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-public-tls-certs\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.023274 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.025053 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-config-data\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.052395 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2lqf\" (UniqueName: \"kubernetes.io/projected/218ec494-3f53-4cb7-8d33-10f7824f4e7c-kube-api-access-x2lqf\") pod \"nova-api-0\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.060288 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.107691 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.579146 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 20:27:57 crc kubenswrapper[4726]: W1123 20:27:57.584356 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b15cde8_6249_48c9_a7f3_984adb91acf1.slice/crio-38ddf454ca40038cd24e949f606fce0f417db1ad06e763f43a7003e22ce5f246 WatchSource:0}: Error finding container 38ddf454ca40038cd24e949f606fce0f417db1ad06e763f43a7003e22ce5f246: Status 404 returned error can't find the container with id 38ddf454ca40038cd24e949f606fce0f417db1ad06e763f43a7003e22ce5f246 Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.615197 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerStarted","Data":"38ddf454ca40038cd24e949f606fce0f417db1ad06e763f43a7003e22ce5f246"} Nov 23 20:27:57 crc kubenswrapper[4726]: I1123 20:27:57.665782 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:27:58 crc kubenswrapper[4726]: I1123 20:27:58.601181 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62224881-e65a-4572-b8ce-f5f538411754" path="/var/lib/kubelet/pods/62224881-e65a-4572-b8ce-f5f538411754/volumes" Nov 23 20:27:58 crc kubenswrapper[4726]: I1123 20:27:58.602628 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4cfd52a-db3c-4f20-89b3-a29921534365" path="/var/lib/kubelet/pods/c4cfd52a-db3c-4f20-89b3-a29921534365/volumes" Nov 23 20:27:58 crc kubenswrapper[4726]: I1123 20:27:58.660313 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"218ec494-3f53-4cb7-8d33-10f7824f4e7c","Type":"ContainerStarted","Data":"bb145f742f63f63cabae77b800d9be5f6f2920fa5ae506866666d0adf546703f"} Nov 23 20:27:58 crc kubenswrapper[4726]: I1123 20:27:58.660365 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"218ec494-3f53-4cb7-8d33-10f7824f4e7c","Type":"ContainerStarted","Data":"db391c59839dd35e243cd6ade30fde50ccb7533c58267fc6474ffe7ad84ea106"} Nov 23 20:27:58 crc kubenswrapper[4726]: I1123 20:27:58.660377 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"218ec494-3f53-4cb7-8d33-10f7824f4e7c","Type":"ContainerStarted","Data":"20483ec9533c66004a4cf111c8a875f1552278e0853e3e8f45cd9c3bca53d073"} Nov 23 20:27:58 crc kubenswrapper[4726]: I1123 20:27:58.667903 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerStarted","Data":"5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5"} Nov 23 20:27:59 crc kubenswrapper[4726]: I1123 20:27:59.676916 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerStarted","Data":"a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d"} Nov 23 20:27:59 crc kubenswrapper[4726]: I1123 20:27:59.677208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerStarted","Data":"862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3"} Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.032043 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.058653 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.058631026 podStartE2EDuration="4.058631026s" podCreationTimestamp="2025-11-23 20:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:27:58.685554195 +0000 UTC m=+1186.834595171" watchObservedRunningTime="2025-11-23 20:28:00.058631026 +0000 UTC m=+1188.207672002" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.111443 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-gfjd8"] Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.111737 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" podUID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" containerName="dnsmasq-dns" containerID="cri-o://b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102" gracePeriod=10 Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.630483 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.696376 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-sb\") pod \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.696460 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-nb\") pod \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.696593 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzkc7\" (UniqueName: \"kubernetes.io/projected/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-kube-api-access-vzkc7\") pod \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.696611 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-config\") pod \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.696767 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-dns-svc\") pod \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\" (UID: \"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613\") " Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.707160 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-kube-api-access-vzkc7" (OuterVolumeSpecName: "kube-api-access-vzkc7") pod "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" (UID: "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613"). InnerVolumeSpecName "kube-api-access-vzkc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.707207 4726 generic.go:334] "Generic (PLEG): container finished" podID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" containerID="b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102" exitCode=0 Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.707245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" event={"ID":"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613","Type":"ContainerDied","Data":"b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102"} Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.707272 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" event={"ID":"c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613","Type":"ContainerDied","Data":"49db783d1234e2fa407df204f7e99d6733f28b4307fe5e1203b55cfce525bdee"} Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.707285 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69494d9f89-gfjd8" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.707290 4726 scope.go:117] "RemoveContainer" containerID="b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.750119 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" (UID: "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.775271 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" (UID: "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.797663 4726 scope.go:117] "RemoveContainer" containerID="93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.798900 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.798930 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzkc7\" (UniqueName: \"kubernetes.io/projected/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-kube-api-access-vzkc7\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.798941 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.808688 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" (UID: "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.808848 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-config" (OuterVolumeSpecName: "config") pod "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" (UID: "c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.831131 4726 scope.go:117] "RemoveContainer" containerID="b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102" Nov 23 20:28:00 crc kubenswrapper[4726]: E1123 20:28:00.833441 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102\": container with ID starting with b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102 not found: ID does not exist" containerID="b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.833499 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102"} err="failed to get container status \"b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102\": rpc error: code = NotFound desc = could not find container \"b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102\": container with ID starting with b68faebc592b65c38f87ab6f5773002e44c3ebf5a00a180c36f932edb80cf102 not found: ID does not exist" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.833536 4726 scope.go:117] "RemoveContainer" containerID="93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494" Nov 23 20:28:00 crc kubenswrapper[4726]: E1123 20:28:00.836420 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494\": container with ID starting with 93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494 not found: ID does not exist" containerID="93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.836457 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494"} err="failed to get container status \"93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494\": rpc error: code = NotFound desc = could not find container \"93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494\": container with ID starting with 93c7ecd4242e5d146960a172317a55fc68a2f5bd7d40d13f4971d7975d4f5494 not found: ID does not exist" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.900630 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:00 crc kubenswrapper[4726]: I1123 20:28:00.901125 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:01 crc kubenswrapper[4726]: I1123 20:28:01.039486 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-gfjd8"] Nov 23 20:28:01 crc kubenswrapper[4726]: I1123 20:28:01.047224 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69494d9f89-gfjd8"] Nov 23 20:28:01 crc kubenswrapper[4726]: I1123 20:28:01.718927 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerStarted","Data":"af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f"} Nov 23 20:28:01 crc kubenswrapper[4726]: I1123 20:28:01.719109 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 20:28:01 crc kubenswrapper[4726]: I1123 20:28:01.750353 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.684053173 podStartE2EDuration="5.750338034s" podCreationTimestamp="2025-11-23 20:27:56 +0000 UTC" firstStartedPulling="2025-11-23 20:27:57.588102429 +0000 UTC m=+1185.737143405" lastFinishedPulling="2025-11-23 20:28:00.65438731 +0000 UTC m=+1188.803428266" observedRunningTime="2025-11-23 20:28:01.746554049 +0000 UTC m=+1189.895595005" watchObservedRunningTime="2025-11-23 20:28:01.750338034 +0000 UTC m=+1189.899378990" Nov 23 20:28:02 crc kubenswrapper[4726]: I1123 20:28:02.600054 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" path="/var/lib/kubelet/pods/c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613/volumes" Nov 23 20:28:02 crc kubenswrapper[4726]: I1123 20:28:02.732159 4726 generic.go:334] "Generic (PLEG): container finished" podID="267f6c6b-3b5d-4e89-bf80-3650b890aed9" containerID="cff0b94346727804ad9dcbf63d396db0137c49417b03f512ba5091ab5bb59225" exitCode=0 Nov 23 20:28:02 crc kubenswrapper[4726]: I1123 20:28:02.732244 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8qvxw" event={"ID":"267f6c6b-3b5d-4e89-bf80-3650b890aed9","Type":"ContainerDied","Data":"cff0b94346727804ad9dcbf63d396db0137c49417b03f512ba5091ab5bb59225"} Nov 23 20:28:03 crc kubenswrapper[4726]: I1123 20:28:03.868023 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 20:28:03 crc kubenswrapper[4726]: I1123 20:28:03.870058 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 20:28:03 crc kubenswrapper[4726]: I1123 20:28:03.879760 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.100147 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.175906 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-config-data\") pod \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.176046 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-scripts\") pod \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.176121 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmjp6\" (UniqueName: \"kubernetes.io/projected/267f6c6b-3b5d-4e89-bf80-3650b890aed9-kube-api-access-jmjp6\") pod \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.176247 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-combined-ca-bundle\") pod \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\" (UID: \"267f6c6b-3b5d-4e89-bf80-3650b890aed9\") " Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.181755 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-scripts" (OuterVolumeSpecName: "scripts") pod "267f6c6b-3b5d-4e89-bf80-3650b890aed9" (UID: "267f6c6b-3b5d-4e89-bf80-3650b890aed9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.182229 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/267f6c6b-3b5d-4e89-bf80-3650b890aed9-kube-api-access-jmjp6" (OuterVolumeSpecName: "kube-api-access-jmjp6") pod "267f6c6b-3b5d-4e89-bf80-3650b890aed9" (UID: "267f6c6b-3b5d-4e89-bf80-3650b890aed9"). InnerVolumeSpecName "kube-api-access-jmjp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.204518 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-config-data" (OuterVolumeSpecName: "config-data") pod "267f6c6b-3b5d-4e89-bf80-3650b890aed9" (UID: "267f6c6b-3b5d-4e89-bf80-3650b890aed9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.208548 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "267f6c6b-3b5d-4e89-bf80-3650b890aed9" (UID: "267f6c6b-3b5d-4e89-bf80-3650b890aed9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.278911 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.278942 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.278952 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmjp6\" (UniqueName: \"kubernetes.io/projected/267f6c6b-3b5d-4e89-bf80-3650b890aed9-kube-api-access-jmjp6\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.278962 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/267f6c6b-3b5d-4e89-bf80-3650b890aed9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.752361 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8qvxw" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.752333 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8qvxw" event={"ID":"267f6c6b-3b5d-4e89-bf80-3650b890aed9","Type":"ContainerDied","Data":"e57ff1d16aaab28cf695f973cda52b180f43fc46e60b1c34b69e33dd449891fc"} Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.752827 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e57ff1d16aaab28cf695f973cda52b180f43fc46e60b1c34b69e33dd449891fc" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.761756 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.963198 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.963437 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a310fd7f-5f37-4428-ba58-a8f4184ce283" containerName="nova-scheduler-scheduler" containerID="cri-o://26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152" gracePeriod=30 Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.978171 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.978478 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerName="nova-api-log" containerID="cri-o://db391c59839dd35e243cd6ade30fde50ccb7533c58267fc6474ffe7ad84ea106" gracePeriod=30 Nov 23 20:28:04 crc kubenswrapper[4726]: I1123 20:28:04.978547 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerName="nova-api-api" containerID="cri-o://bb145f742f63f63cabae77b800d9be5f6f2920fa5ae506866666d0adf546703f" gracePeriod=30 Nov 23 20:28:05 crc kubenswrapper[4726]: I1123 20:28:05.015738 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:28:05 crc kubenswrapper[4726]: I1123 20:28:05.766022 4726 generic.go:334] "Generic (PLEG): container finished" podID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerID="bb145f742f63f63cabae77b800d9be5f6f2920fa5ae506866666d0adf546703f" exitCode=0 Nov 23 20:28:05 crc kubenswrapper[4726]: I1123 20:28:05.766356 4726 generic.go:334] "Generic (PLEG): container finished" podID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerID="db391c59839dd35e243cd6ade30fde50ccb7533c58267fc6474ffe7ad84ea106" exitCode=143 Nov 23 20:28:05 crc kubenswrapper[4726]: I1123 20:28:05.767216 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"218ec494-3f53-4cb7-8d33-10f7824f4e7c","Type":"ContainerDied","Data":"bb145f742f63f63cabae77b800d9be5f6f2920fa5ae506866666d0adf546703f"} Nov 23 20:28:05 crc kubenswrapper[4726]: I1123 20:28:05.767251 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"218ec494-3f53-4cb7-8d33-10f7824f4e7c","Type":"ContainerDied","Data":"db391c59839dd35e243cd6ade30fde50ccb7533c58267fc6474ffe7ad84ea106"} Nov 23 20:28:05 crc kubenswrapper[4726]: I1123 20:28:05.876369 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.017901 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2lqf\" (UniqueName: \"kubernetes.io/projected/218ec494-3f53-4cb7-8d33-10f7824f4e7c-kube-api-access-x2lqf\") pod \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.017967 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-config-data\") pod \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.017986 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-public-tls-certs\") pod \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.018124 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-internal-tls-certs\") pod \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.018174 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218ec494-3f53-4cb7-8d33-10f7824f4e7c-logs\") pod \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.018196 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-combined-ca-bundle\") pod \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\" (UID: \"218ec494-3f53-4cb7-8d33-10f7824f4e7c\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.022529 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/218ec494-3f53-4cb7-8d33-10f7824f4e7c-logs" (OuterVolumeSpecName: "logs") pod "218ec494-3f53-4cb7-8d33-10f7824f4e7c" (UID: "218ec494-3f53-4cb7-8d33-10f7824f4e7c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.030232 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/218ec494-3f53-4cb7-8d33-10f7824f4e7c-kube-api-access-x2lqf" (OuterVolumeSpecName: "kube-api-access-x2lqf") pod "218ec494-3f53-4cb7-8d33-10f7824f4e7c" (UID: "218ec494-3f53-4cb7-8d33-10f7824f4e7c"). InnerVolumeSpecName "kube-api-access-x2lqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.055611 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-config-data" (OuterVolumeSpecName: "config-data") pod "218ec494-3f53-4cb7-8d33-10f7824f4e7c" (UID: "218ec494-3f53-4cb7-8d33-10f7824f4e7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.074291 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "218ec494-3f53-4cb7-8d33-10f7824f4e7c" (UID: "218ec494-3f53-4cb7-8d33-10f7824f4e7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.076182 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "218ec494-3f53-4cb7-8d33-10f7824f4e7c" (UID: "218ec494-3f53-4cb7-8d33-10f7824f4e7c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.081662 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "218ec494-3f53-4cb7-8d33-10f7824f4e7c" (UID: "218ec494-3f53-4cb7-8d33-10f7824f4e7c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.126431 4726 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.126470 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/218ec494-3f53-4cb7-8d33-10f7824f4e7c-logs\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.126479 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.126488 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2lqf\" (UniqueName: \"kubernetes.io/projected/218ec494-3f53-4cb7-8d33-10f7824f4e7c-kube-api-access-x2lqf\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.126498 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.126506 4726 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/218ec494-3f53-4cb7-8d33-10f7824f4e7c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.353911 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.432115 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-config-data\") pod \"a310fd7f-5f37-4428-ba58-a8f4184ce283\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.432159 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-combined-ca-bundle\") pod \"a310fd7f-5f37-4428-ba58-a8f4184ce283\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.432225 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwvhc\" (UniqueName: \"kubernetes.io/projected/a310fd7f-5f37-4428-ba58-a8f4184ce283-kube-api-access-lwvhc\") pod \"a310fd7f-5f37-4428-ba58-a8f4184ce283\" (UID: \"a310fd7f-5f37-4428-ba58-a8f4184ce283\") " Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.448658 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a310fd7f-5f37-4428-ba58-a8f4184ce283-kube-api-access-lwvhc" (OuterVolumeSpecName: "kube-api-access-lwvhc") pod "a310fd7f-5f37-4428-ba58-a8f4184ce283" (UID: "a310fd7f-5f37-4428-ba58-a8f4184ce283"). InnerVolumeSpecName "kube-api-access-lwvhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.457545 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a310fd7f-5f37-4428-ba58-a8f4184ce283" (UID: "a310fd7f-5f37-4428-ba58-a8f4184ce283"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.477943 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-config-data" (OuterVolumeSpecName: "config-data") pod "a310fd7f-5f37-4428-ba58-a8f4184ce283" (UID: "a310fd7f-5f37-4428-ba58-a8f4184ce283"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.534445 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.534478 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a310fd7f-5f37-4428-ba58-a8f4184ce283-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.534493 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwvhc\" (UniqueName: \"kubernetes.io/projected/a310fd7f-5f37-4428-ba58-a8f4184ce283-kube-api-access-lwvhc\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.776541 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"218ec494-3f53-4cb7-8d33-10f7824f4e7c","Type":"ContainerDied","Data":"20483ec9533c66004a4cf111c8a875f1552278e0853e3e8f45cd9c3bca53d073"} Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.776595 4726 scope.go:117] "RemoveContainer" containerID="bb145f742f63f63cabae77b800d9be5f6f2920fa5ae506866666d0adf546703f" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.776599 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.778561 4726 generic.go:334] "Generic (PLEG): container finished" podID="a310fd7f-5f37-4428-ba58-a8f4184ce283" containerID="26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152" exitCode=0 Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.778770 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-log" containerID="cri-o://9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe" gracePeriod=30 Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.778809 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a310fd7f-5f37-4428-ba58-a8f4184ce283","Type":"ContainerDied","Data":"26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152"} Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.778859 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a310fd7f-5f37-4428-ba58-a8f4184ce283","Type":"ContainerDied","Data":"f9f35cb839998f0bc61dd01ad027746eca9efb1922a97a280d054e04a5998667"} Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.778819 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.778890 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-metadata" containerID="cri-o://ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a" gracePeriod=30 Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.798333 4726 scope.go:117] "RemoveContainer" containerID="db391c59839dd35e243cd6ade30fde50ccb7533c58267fc6474ffe7ad84ea106" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.812925 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.818034 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.820324 4726 scope.go:117] "RemoveContainer" containerID="26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.850829 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 23 20:28:06 crc kubenswrapper[4726]: E1123 20:28:06.851282 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerName="nova-api-log" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851301 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerName="nova-api-log" Nov 23 20:28:06 crc kubenswrapper[4726]: E1123 20:28:06.851314 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" containerName="init" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851320 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" containerName="init" Nov 23 20:28:06 crc kubenswrapper[4726]: E1123 20:28:06.851347 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" containerName="dnsmasq-dns" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851354 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" containerName="dnsmasq-dns" Nov 23 20:28:06 crc kubenswrapper[4726]: E1123 20:28:06.851370 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerName="nova-api-api" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851376 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerName="nova-api-api" Nov 23 20:28:06 crc kubenswrapper[4726]: E1123 20:28:06.851388 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267f6c6b-3b5d-4e89-bf80-3650b890aed9" containerName="nova-manage" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851395 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="267f6c6b-3b5d-4e89-bf80-3650b890aed9" containerName="nova-manage" Nov 23 20:28:06 crc kubenswrapper[4726]: E1123 20:28:06.851408 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a310fd7f-5f37-4428-ba58-a8f4184ce283" containerName="nova-scheduler-scheduler" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851414 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a310fd7f-5f37-4428-ba58-a8f4184ce283" containerName="nova-scheduler-scheduler" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851744 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ab62d2-07f5-4a6b-95aa-e0f52f9b0613" containerName="dnsmasq-dns" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851754 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerName="nova-api-api" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851762 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a310fd7f-5f37-4428-ba58-a8f4184ce283" containerName="nova-scheduler-scheduler" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851775 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" containerName="nova-api-log" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.851782 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="267f6c6b-3b5d-4e89-bf80-3650b890aed9" containerName="nova-manage" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.852827 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.853771 4726 scope.go:117] "RemoveContainer" containerID="26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152" Nov 23 20:28:06 crc kubenswrapper[4726]: E1123 20:28:06.854504 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152\": container with ID starting with 26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152 not found: ID does not exist" containerID="26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.854609 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152"} err="failed to get container status \"26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152\": rpc error: code = NotFound desc = could not find container \"26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152\": container with ID starting with 26e896f957e3a2a33abe7cc5bac751d06bf3479cd612e679e91b88c4ae15a152 not found: ID does not exist" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.855013 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.861484 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.868999 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.873426 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.905021 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.905091 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.906479 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.909937 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.918135 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.941849 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.941926 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzmmh\" (UniqueName: \"kubernetes.io/projected/29188283-47f2-45d2-b6ab-d3cc9b97faff-kube-api-access-xzmmh\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.941971 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-public-tls-certs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.942000 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29188283-47f2-45d2-b6ab-d3cc9b97faff-logs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.942026 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-config-data\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.942062 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:06 crc kubenswrapper[4726]: I1123 20:28:06.943471 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044060 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-public-tls-certs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044113 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e324e799-3ac3-4e70-a035-822dff0d5972-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044137 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zgfz\" (UniqueName: \"kubernetes.io/projected/e324e799-3ac3-4e70-a035-822dff0d5972-kube-api-access-6zgfz\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044170 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29188283-47f2-45d2-b6ab-d3cc9b97faff-logs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-config-data\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044237 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044336 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e324e799-3ac3-4e70-a035-822dff0d5972-config-data\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044646 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.044698 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzmmh\" (UniqueName: \"kubernetes.io/projected/29188283-47f2-45d2-b6ab-d3cc9b97faff-kube-api-access-xzmmh\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.045088 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29188283-47f2-45d2-b6ab-d3cc9b97faff-logs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.050847 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.052386 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.052582 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-public-tls-certs\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.052858 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29188283-47f2-45d2-b6ab-d3cc9b97faff-config-data\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.061031 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzmmh\" (UniqueName: \"kubernetes.io/projected/29188283-47f2-45d2-b6ab-d3cc9b97faff-kube-api-access-xzmmh\") pod \"nova-api-0\" (UID: \"29188283-47f2-45d2-b6ab-d3cc9b97faff\") " pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.146566 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e324e799-3ac3-4e70-a035-822dff0d5972-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.146857 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zgfz\" (UniqueName: \"kubernetes.io/projected/e324e799-3ac3-4e70-a035-822dff0d5972-kube-api-access-6zgfz\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.147018 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e324e799-3ac3-4e70-a035-822dff0d5972-config-data\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.150601 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e324e799-3ac3-4e70-a035-822dff0d5972-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.152400 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e324e799-3ac3-4e70-a035-822dff0d5972-config-data\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.175466 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zgfz\" (UniqueName: \"kubernetes.io/projected/e324e799-3ac3-4e70-a035-822dff0d5972-kube-api-access-6zgfz\") pod \"nova-scheduler-0\" (UID: \"e324e799-3ac3-4e70-a035-822dff0d5972\") " pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.226173 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.232390 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.774145 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.804318 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.806136 4726 generic.go:334] "Generic (PLEG): container finished" podID="364f0735-33a2-4a68-88ac-86842a3a5738" containerID="9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe" exitCode=143 Nov 23 20:28:07 crc kubenswrapper[4726]: I1123 20:28:07.806194 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"364f0735-33a2-4a68-88ac-86842a3a5738","Type":"ContainerDied","Data":"9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe"} Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.613427 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="218ec494-3f53-4cb7-8d33-10f7824f4e7c" path="/var/lib/kubelet/pods/218ec494-3f53-4cb7-8d33-10f7824f4e7c/volumes" Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.615532 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a310fd7f-5f37-4428-ba58-a8f4184ce283" path="/var/lib/kubelet/pods/a310fd7f-5f37-4428-ba58-a8f4184ce283/volumes" Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.824347 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29188283-47f2-45d2-b6ab-d3cc9b97faff","Type":"ContainerStarted","Data":"5b75700f1edd6167d7a18d738f22337902f9c6251c82730a58c860f20fb72a48"} Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.824430 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29188283-47f2-45d2-b6ab-d3cc9b97faff","Type":"ContainerStarted","Data":"98b987a544cb1588cdaad3df94bcb58c2b382bc011e6978125d39542edd66dcb"} Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.824474 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29188283-47f2-45d2-b6ab-d3cc9b97faff","Type":"ContainerStarted","Data":"5ea79587888e77b8dee603dbff25d434a435aab3654eb7860c49c859d2573f5f"} Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.826530 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e324e799-3ac3-4e70-a035-822dff0d5972","Type":"ContainerStarted","Data":"c778492ed82067b9e06d264a7f1fd24601f180bb8b27ea0f46bc15782ec986b5"} Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.826584 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e324e799-3ac3-4e70-a035-822dff0d5972","Type":"ContainerStarted","Data":"31ab8ec46992c04a956fa6097a11c7b58eb16e5fea86976f814bd13fba0764d4"} Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.863629 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.86359946 podStartE2EDuration="2.86359946s" podCreationTimestamp="2025-11-23 20:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:28:08.852448382 +0000 UTC m=+1197.001489358" watchObservedRunningTime="2025-11-23 20:28:08.86359946 +0000 UTC m=+1197.012640426" Nov 23 20:28:08 crc kubenswrapper[4726]: I1123 20:28:08.889705 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.889677442 podStartE2EDuration="2.889677442s" podCreationTimestamp="2025-11-23 20:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:28:08.877389192 +0000 UTC m=+1197.026430218" watchObservedRunningTime="2025-11-23 20:28:08.889677442 +0000 UTC m=+1197.038718408" Nov 23 20:28:09 crc kubenswrapper[4726]: I1123 20:28:09.970121 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": read tcp 10.217.0.2:58000->10.217.0.181:8775: read: connection reset by peer" Nov 23 20:28:09 crc kubenswrapper[4726]: I1123 20:28:09.970193 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.181:8775/\": read tcp 10.217.0.2:58004->10.217.0.181:8775: read: connection reset by peer" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.355902 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.424723 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwj7c\" (UniqueName: \"kubernetes.io/projected/364f0735-33a2-4a68-88ac-86842a3a5738-kube-api-access-zwj7c\") pod \"364f0735-33a2-4a68-88ac-86842a3a5738\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.425015 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-config-data\") pod \"364f0735-33a2-4a68-88ac-86842a3a5738\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.425145 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-combined-ca-bundle\") pod \"364f0735-33a2-4a68-88ac-86842a3a5738\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.425878 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-nova-metadata-tls-certs\") pod \"364f0735-33a2-4a68-88ac-86842a3a5738\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.425965 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364f0735-33a2-4a68-88ac-86842a3a5738-logs\") pod \"364f0735-33a2-4a68-88ac-86842a3a5738\" (UID: \"364f0735-33a2-4a68-88ac-86842a3a5738\") " Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.426786 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/364f0735-33a2-4a68-88ac-86842a3a5738-logs" (OuterVolumeSpecName: "logs") pod "364f0735-33a2-4a68-88ac-86842a3a5738" (UID: "364f0735-33a2-4a68-88ac-86842a3a5738"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.448262 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364f0735-33a2-4a68-88ac-86842a3a5738-kube-api-access-zwj7c" (OuterVolumeSpecName: "kube-api-access-zwj7c") pod "364f0735-33a2-4a68-88ac-86842a3a5738" (UID: "364f0735-33a2-4a68-88ac-86842a3a5738"). InnerVolumeSpecName "kube-api-access-zwj7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.461648 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "364f0735-33a2-4a68-88ac-86842a3a5738" (UID: "364f0735-33a2-4a68-88ac-86842a3a5738"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.466979 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-config-data" (OuterVolumeSpecName: "config-data") pod "364f0735-33a2-4a68-88ac-86842a3a5738" (UID: "364f0735-33a2-4a68-88ac-86842a3a5738"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.497018 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "364f0735-33a2-4a68-88ac-86842a3a5738" (UID: "364f0735-33a2-4a68-88ac-86842a3a5738"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.528370 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.528404 4726 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.528416 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/364f0735-33a2-4a68-88ac-86842a3a5738-logs\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.528426 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwj7c\" (UniqueName: \"kubernetes.io/projected/364f0735-33a2-4a68-88ac-86842a3a5738-kube-api-access-zwj7c\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.528435 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364f0735-33a2-4a68-88ac-86842a3a5738-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.847480 4726 generic.go:334] "Generic (PLEG): container finished" podID="364f0735-33a2-4a68-88ac-86842a3a5738" containerID="ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a" exitCode=0 Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.847569 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.847582 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"364f0735-33a2-4a68-88ac-86842a3a5738","Type":"ContainerDied","Data":"ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a"} Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.849641 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"364f0735-33a2-4a68-88ac-86842a3a5738","Type":"ContainerDied","Data":"bdf1c953be13c061d6221e533a8c02887a55cc982c00228598d52f4c517192dc"} Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.849837 4726 scope.go:117] "RemoveContainer" containerID="ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.887444 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.887628 4726 scope.go:117] "RemoveContainer" containerID="9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.896525 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.913927 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.914281 4726 scope.go:117] "RemoveContainer" containerID="ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a" Nov 23 20:28:10 crc kubenswrapper[4726]: E1123 20:28:10.914789 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a\": container with ID starting with ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a not found: ID does not exist" containerID="ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.914827 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a"} err="failed to get container status \"ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a\": rpc error: code = NotFound desc = could not find container \"ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a\": container with ID starting with ac58e966f2488183778588c2fcbdef72cea210fd353230de7b7166d1d16c737a not found: ID does not exist" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.914856 4726 scope.go:117] "RemoveContainer" containerID="9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe" Nov 23 20:28:10 crc kubenswrapper[4726]: E1123 20:28:10.914918 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-log" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.915130 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-log" Nov 23 20:28:10 crc kubenswrapper[4726]: E1123 20:28:10.915191 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-metadata" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.915200 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-metadata" Nov 23 20:28:10 crc kubenswrapper[4726]: E1123 20:28:10.915102 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe\": container with ID starting with 9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe not found: ID does not exist" containerID="9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.915461 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe"} err="failed to get container status \"9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe\": rpc error: code = NotFound desc = could not find container \"9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe\": container with ID starting with 9110bd442f6ac1e6cedbdd245f4719c78f27c8275b41e42b0c42ebf3cad42afe not found: ID does not exist" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.915436 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-log" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.915532 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" containerName="nova-metadata-metadata" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.917401 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.919398 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.931789 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 23 20:28:10 crc kubenswrapper[4726]: I1123 20:28:10.932747 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.037656 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcm2q\" (UniqueName: \"kubernetes.io/projected/7e68d296-e4d0-4462-9b15-1700400d38fa-kube-api-access-rcm2q\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.037816 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e68d296-e4d0-4462-9b15-1700400d38fa-logs\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.037913 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.037946 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-config-data\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.038017 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.139602 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.139671 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-config-data\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.139707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.139731 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcm2q\" (UniqueName: \"kubernetes.io/projected/7e68d296-e4d0-4462-9b15-1700400d38fa-kube-api-access-rcm2q\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.139803 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e68d296-e4d0-4462-9b15-1700400d38fa-logs\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.140224 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e68d296-e4d0-4462-9b15-1700400d38fa-logs\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.143981 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-config-data\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.145235 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.146543 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e68d296-e4d0-4462-9b15-1700400d38fa-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.169590 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcm2q\" (UniqueName: \"kubernetes.io/projected/7e68d296-e4d0-4462-9b15-1700400d38fa-kube-api-access-rcm2q\") pod \"nova-metadata-0\" (UID: \"7e68d296-e4d0-4462-9b15-1700400d38fa\") " pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.263961 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 23 20:28:11 crc kubenswrapper[4726]: W1123 20:28:11.787835 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e68d296_e4d0_4462_9b15_1700400d38fa.slice/crio-bee5577b74c8adc290cf6c8fb1991980d78dec10379471f25d584d9319611c1e WatchSource:0}: Error finding container bee5577b74c8adc290cf6c8fb1991980d78dec10379471f25d584d9319611c1e: Status 404 returned error can't find the container with id bee5577b74c8adc290cf6c8fb1991980d78dec10379471f25d584d9319611c1e Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.788215 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 23 20:28:11 crc kubenswrapper[4726]: I1123 20:28:11.860763 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e68d296-e4d0-4462-9b15-1700400d38fa","Type":"ContainerStarted","Data":"bee5577b74c8adc290cf6c8fb1991980d78dec10379471f25d584d9319611c1e"} Nov 23 20:28:12 crc kubenswrapper[4726]: I1123 20:28:12.232560 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 23 20:28:12 crc kubenswrapper[4726]: I1123 20:28:12.607565 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="364f0735-33a2-4a68-88ac-86842a3a5738" path="/var/lib/kubelet/pods/364f0735-33a2-4a68-88ac-86842a3a5738/volumes" Nov 23 20:28:12 crc kubenswrapper[4726]: I1123 20:28:12.875923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e68d296-e4d0-4462-9b15-1700400d38fa","Type":"ContainerStarted","Data":"8a15c124c597d43a89846b15eb5a80e94966e4bac9713240c5f541858ba206b0"} Nov 23 20:28:12 crc kubenswrapper[4726]: I1123 20:28:12.876324 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7e68d296-e4d0-4462-9b15-1700400d38fa","Type":"ContainerStarted","Data":"63dda9a223a5b5fea8fe27bd3d63d950294e3a4921f73771ef5988ff81592edf"} Nov 23 20:28:12 crc kubenswrapper[4726]: I1123 20:28:12.915510 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.915477641 podStartE2EDuration="2.915477641s" podCreationTimestamp="2025-11-23 20:28:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:28:12.906162824 +0000 UTC m=+1201.055203840" watchObservedRunningTime="2025-11-23 20:28:12.915477641 +0000 UTC m=+1201.064518627" Nov 23 20:28:16 crc kubenswrapper[4726]: I1123 20:28:16.208136 4726 scope.go:117] "RemoveContainer" containerID="443f2466af374ec6ab66aba4901ffb04089b458155fc45edc59498224429efa5" Nov 23 20:28:16 crc kubenswrapper[4726]: I1123 20:28:16.250702 4726 scope.go:117] "RemoveContainer" containerID="1e75fbabd23d794bc28c322fb6306f4bb81e3e3dee5a3c1f797b988adddaf521" Nov 23 20:28:16 crc kubenswrapper[4726]: I1123 20:28:16.264634 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 20:28:16 crc kubenswrapper[4726]: I1123 20:28:16.269068 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 23 20:28:16 crc kubenswrapper[4726]: I1123 20:28:16.276314 4726 scope.go:117] "RemoveContainer" containerID="f1015cd3957770465a824c78204b4fbd3df96dab1f7ec6c3cd4ac58cba93c856" Nov 23 20:28:17 crc kubenswrapper[4726]: I1123 20:28:17.226291 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 20:28:17 crc kubenswrapper[4726]: I1123 20:28:17.226712 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 23 20:28:17 crc kubenswrapper[4726]: I1123 20:28:17.232522 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 23 20:28:17 crc kubenswrapper[4726]: I1123 20:28:17.278404 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 23 20:28:17 crc kubenswrapper[4726]: I1123 20:28:17.990588 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 23 20:28:18 crc kubenswrapper[4726]: I1123 20:28:18.241016 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="29188283-47f2-45d2-b6ab-d3cc9b97faff" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.186:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 20:28:18 crc kubenswrapper[4726]: I1123 20:28:18.241038 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="29188283-47f2-45d2-b6ab-d3cc9b97faff" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 23 20:28:21 crc kubenswrapper[4726]: I1123 20:28:21.265721 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 20:28:21 crc kubenswrapper[4726]: I1123 20:28:21.267165 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 23 20:28:22 crc kubenswrapper[4726]: I1123 20:28:22.278055 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7e68d296-e4d0-4462-9b15-1700400d38fa" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.188:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 20:28:22 crc kubenswrapper[4726]: I1123 20:28:22.278120 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7e68d296-e4d0-4462-9b15-1700400d38fa" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.188:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 23 20:28:27 crc kubenswrapper[4726]: I1123 20:28:27.112351 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 20:28:27 crc kubenswrapper[4726]: I1123 20:28:27.236845 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 20:28:27 crc kubenswrapper[4726]: I1123 20:28:27.237455 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 20:28:27 crc kubenswrapper[4726]: I1123 20:28:27.244584 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 20:28:27 crc kubenswrapper[4726]: I1123 20:28:27.268208 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 23 20:28:28 crc kubenswrapper[4726]: I1123 20:28:28.057519 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 23 20:28:28 crc kubenswrapper[4726]: I1123 20:28:28.072720 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 23 20:28:31 crc kubenswrapper[4726]: I1123 20:28:31.269273 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 20:28:31 crc kubenswrapper[4726]: I1123 20:28:31.269607 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 23 20:28:31 crc kubenswrapper[4726]: I1123 20:28:31.276359 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 20:28:31 crc kubenswrapper[4726]: I1123 20:28:31.278168 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 23 20:28:39 crc kubenswrapper[4726]: I1123 20:28:39.310183 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:28:40 crc kubenswrapper[4726]: I1123 20:28:40.766177 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:28:44 crc kubenswrapper[4726]: I1123 20:28:44.163398 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="c1879312-c440-415f-a376-4ab58b99e21d" containerName="rabbitmq" containerID="cri-o://1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18" gracePeriod=604796 Nov 23 20:28:44 crc kubenswrapper[4726]: I1123 20:28:44.815415 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="c1879312-c440-415f-a376-4ab58b99e21d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Nov 23 20:28:45 crc kubenswrapper[4726]: I1123 20:28:45.850175 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerName="rabbitmq" containerID="cri-o://2471650b3dd77658e1c60d02e48d6b6a1f06b40fb3b935a5ed1903e883f8b86c" gracePeriod=604795 Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.722547 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854408 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x4ht\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-kube-api-access-7x4ht\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854517 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1879312-c440-415f-a376-4ab58b99e21d-erlang-cookie-secret\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854570 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1879312-c440-415f-a376-4ab58b99e21d-pod-info\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854597 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-confd\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854623 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-erlang-cookie\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854659 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-config-data\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854691 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-plugins-conf\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854760 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-plugins\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854819 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854856 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-server-conf\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.854932 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-tls\") pod \"c1879312-c440-415f-a376-4ab58b99e21d\" (UID: \"c1879312-c440-415f-a376-4ab58b99e21d\") " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.855622 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.855998 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.863366 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.863462 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c1879312-c440-415f-a376-4ab58b99e21d-pod-info" (OuterVolumeSpecName: "pod-info") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.868312 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-kube-api-access-7x4ht" (OuterVolumeSpecName: "kube-api-access-7x4ht") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "kube-api-access-7x4ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.871004 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1879312-c440-415f-a376-4ab58b99e21d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.871021 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.871230 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.896829 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-config-data" (OuterVolumeSpecName: "config-data") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.940764 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-server-conf" (OuterVolumeSpecName: "server-conf") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957296 4726 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1879312-c440-415f-a376-4ab58b99e21d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957326 4726 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1879312-c440-415f-a376-4ab58b99e21d-pod-info\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957336 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957347 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957356 4726 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957365 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957401 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957410 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957420 4726 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1879312-c440-415f-a376-4ab58b99e21d-server-conf\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.957429 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x4ht\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-kube-api-access-7x4ht\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.988150 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 23 20:28:50 crc kubenswrapper[4726]: I1123 20:28:50.994787 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c1879312-c440-415f-a376-4ab58b99e21d" (UID: "c1879312-c440-415f-a376-4ab58b99e21d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.059546 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.059583 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1879312-c440-415f-a376-4ab58b99e21d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.256999 4726 generic.go:334] "Generic (PLEG): container finished" podID="c1879312-c440-415f-a376-4ab58b99e21d" containerID="1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18" exitCode=0 Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.257043 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1879312-c440-415f-a376-4ab58b99e21d","Type":"ContainerDied","Data":"1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18"} Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.257070 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1879312-c440-415f-a376-4ab58b99e21d","Type":"ContainerDied","Data":"5027f5618fa01a06cafc6319af10587ee946bc63fb08f7432605af486edad51c"} Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.257088 4726 scope.go:117] "RemoveContainer" containerID="1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.257217 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.278851 4726 scope.go:117] "RemoveContainer" containerID="2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.294787 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.306327 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.306761 4726 scope.go:117] "RemoveContainer" containerID="1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18" Nov 23 20:28:51 crc kubenswrapper[4726]: E1123 20:28:51.307251 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18\": container with ID starting with 1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18 not found: ID does not exist" containerID="1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.307289 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18"} err="failed to get container status \"1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18\": rpc error: code = NotFound desc = could not find container \"1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18\": container with ID starting with 1ef18cca599891b4e4422a3ada15f98b02061f5feb5f1a80dd84fa02139c9d18 not found: ID does not exist" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.307316 4726 scope.go:117] "RemoveContainer" containerID="2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f" Nov 23 20:28:51 crc kubenswrapper[4726]: E1123 20:28:51.307571 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f\": container with ID starting with 2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f not found: ID does not exist" containerID="2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.307596 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f"} err="failed to get container status \"2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f\": rpc error: code = NotFound desc = could not find container \"2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f\": container with ID starting with 2bf43dfaf4f91279395b5f46b1a0e97138db15b7b6dfefd5bd598f329375c38f not found: ID does not exist" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.331006 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:28:51 crc kubenswrapper[4726]: E1123 20:28:51.331378 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1879312-c440-415f-a376-4ab58b99e21d" containerName="setup-container" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.331395 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1879312-c440-415f-a376-4ab58b99e21d" containerName="setup-container" Nov 23 20:28:51 crc kubenswrapper[4726]: E1123 20:28:51.331415 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1879312-c440-415f-a376-4ab58b99e21d" containerName="rabbitmq" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.331422 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1879312-c440-415f-a376-4ab58b99e21d" containerName="rabbitmq" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.331582 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1879312-c440-415f-a376-4ab58b99e21d" containerName="rabbitmq" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.332490 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.335938 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.335941 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.336461 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.336834 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.338314 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-kmlks" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.340105 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.341113 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.361491 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467139 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467196 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467354 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f962af44-2b42-4924-81b8-872c96335906-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467498 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467560 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467622 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-config-data\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467691 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f962af44-2b42-4924-81b8-872c96335906-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467798 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d976\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-kube-api-access-9d976\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467825 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467848 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.467889 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569483 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-config-data\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569762 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f962af44-2b42-4924-81b8-872c96335906-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569808 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d976\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-kube-api-access-9d976\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569825 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569844 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569925 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569944 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.569976 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f962af44-2b42-4924-81b8-872c96335906-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.570012 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.570165 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.579782 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.579983 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.580250 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.580324 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-config-data\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.580454 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f962af44-2b42-4924-81b8-872c96335906-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.595710 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f962af44-2b42-4924-81b8-872c96335906-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.596716 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f962af44-2b42-4924-81b8-872c96335906-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.598537 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.602713 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.614658 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.620109 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d976\" (UniqueName: \"kubernetes.io/projected/f962af44-2b42-4924-81b8-872c96335906-kube-api-access-9d976\") pod \"rabbitmq-server-0\" (UID: \"f962af44-2b42-4924-81b8-872c96335906\") " pod="openstack/rabbitmq-server-0" Nov 23 20:28:51 crc kubenswrapper[4726]: I1123 20:28:51.651283 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.265558 4726 generic.go:334] "Generic (PLEG): container finished" podID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerID="2471650b3dd77658e1c60d02e48d6b6a1f06b40fb3b935a5ed1903e883f8b86c" exitCode=0 Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.265937 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6","Type":"ContainerDied","Data":"2471650b3dd77658e1c60d02e48d6b6a1f06b40fb3b935a5ed1903e883f8b86c"} Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.309112 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 23 20:28:52 crc kubenswrapper[4726]: W1123 20:28:52.322041 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf962af44_2b42_4924_81b8_872c96335906.slice/crio-4ef7e2bb3a3b8baef434c8f84b6620397abdb4d247cfe8eb20e903288a92f990 WatchSource:0}: Error finding container 4ef7e2bb3a3b8baef434c8f84b6620397abdb4d247cfe8eb20e903288a92f990: Status 404 returned error can't find the container with id 4ef7e2bb3a3b8baef434c8f84b6620397abdb4d247cfe8eb20e903288a92f990 Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.352814 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518501 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-server-conf\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518547 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-confd\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518608 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-erlang-cookie-secret\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518643 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-erlang-cookie\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518698 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518724 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-tls\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518756 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-pod-info\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518777 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-config-data\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518806 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-plugins\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518833 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjwjq\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-kube-api-access-hjwjq\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.518865 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-plugins-conf\") pod \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\" (UID: \"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6\") " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.519737 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.520958 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.526584 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.527616 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.534324 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-pod-info" (OuterVolumeSpecName: "pod-info") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.535084 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-kube-api-access-hjwjq" (OuterVolumeSpecName: "kube-api-access-hjwjq") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "kube-api-access-hjwjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.538080 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.541325 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.571606 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-config-data" (OuterVolumeSpecName: "config-data") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.594615 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-server-conf" (OuterVolumeSpecName: "server-conf") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.611027 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1879312-c440-415f-a376-4ab58b99e21d" path="/var/lib/kubelet/pods/c1879312-c440-415f-a376-4ab58b99e21d/volumes" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.623979 4726 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-server-conf\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624008 4726 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624018 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624045 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624055 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624063 4726 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-pod-info\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624072 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624080 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624092 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjwjq\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-kube-api-access-hjwjq\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.624102 4726 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.659334 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.674141 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" (UID: "c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.725818 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:52 crc kubenswrapper[4726]: I1123 20:28:52.725855 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.281432 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.282955 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6","Type":"ContainerDied","Data":"c2f9f0dd41e5163141b323218ea9ec9de39e4bed74ef0d6efd23b88f65bbf864"} Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.283274 4726 scope.go:117] "RemoveContainer" containerID="2471650b3dd77658e1c60d02e48d6b6a1f06b40fb3b935a5ed1903e883f8b86c" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.284232 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f962af44-2b42-4924-81b8-872c96335906","Type":"ContainerStarted","Data":"4ef7e2bb3a3b8baef434c8f84b6620397abdb4d247cfe8eb20e903288a92f990"} Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.328752 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.332126 4726 scope.go:117] "RemoveContainer" containerID="23e09746982c7ccb78dd8daed1438ea3180103f7bbf3840ead65e6402c29f107" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.338107 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.364447 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:28:53 crc kubenswrapper[4726]: E1123 20:28:53.364817 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerName="setup-container" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.364833 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerName="setup-container" Nov 23 20:28:53 crc kubenswrapper[4726]: E1123 20:28:53.364846 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerName="rabbitmq" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.364852 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerName="rabbitmq" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.365304 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" containerName="rabbitmq" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.366338 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.369137 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.369617 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.370031 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.370158 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.370266 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-24tl5" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.370411 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.376328 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.396311 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:28:53 crc kubenswrapper[4726]: E1123 20:28:53.438238 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8d94f6c_c1d7_48fe_b7fa_a540d0aae0f6.slice/crio-c2f9f0dd41e5163141b323218ea9ec9de39e4bed74ef0d6efd23b88f65bbf864\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8d94f6c_c1d7_48fe_b7fa_a540d0aae0f6.slice\": RecentStats: unable to find data in memory cache]" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443331 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443381 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443400 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0883d1c4-a52b-4221-a651-8ba8c5463dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443433 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443450 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443559 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443580 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443611 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443629 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnzm7\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-kube-api-access-gnzm7\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443660 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0883d1c4-a52b-4221-a651-8ba8c5463dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.443709 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.544803 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.544854 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.544886 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0883d1c4-a52b-4221-a651-8ba8c5463dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.544917 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.544934 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.544982 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.545001 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.545027 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.545043 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnzm7\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-kube-api-access-gnzm7\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.545070 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0883d1c4-a52b-4221-a651-8ba8c5463dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.545107 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.545248 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.545371 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.545836 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.546101 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.546393 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.546984 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0883d1c4-a52b-4221-a651-8ba8c5463dd6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.552764 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0883d1c4-a52b-4221-a651-8ba8c5463dd6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.587618 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.588021 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.588212 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0883d1c4-a52b-4221-a651-8ba8c5463dd6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.588428 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnzm7\" (UniqueName: \"kubernetes.io/projected/0883d1c4-a52b-4221-a651-8ba8c5463dd6-kube-api-access-gnzm7\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.627929 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0883d1c4-a52b-4221-a651-8ba8c5463dd6\") " pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:53 crc kubenswrapper[4726]: I1123 20:28:53.778322 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:28:54 crc kubenswrapper[4726]: I1123 20:28:54.296896 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f962af44-2b42-4924-81b8-872c96335906","Type":"ContainerStarted","Data":"908c9a60dbbc81790112cf7531d732f69abb86cee42b37a5f3087b22aea3b991"} Nov 23 20:28:54 crc kubenswrapper[4726]: I1123 20:28:54.316441 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 23 20:28:54 crc kubenswrapper[4726]: I1123 20:28:54.606167 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6" path="/var/lib/kubelet/pods/c8d94f6c-c1d7-48fe-b7fa-a540d0aae0f6/volumes" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.310014 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0883d1c4-a52b-4221-a651-8ba8c5463dd6","Type":"ContainerStarted","Data":"36744f2cf3ecb6c48b412cf0a37aa8b34a484fa317f779d0dc65073234716cc7"} Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.740785 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568675b579-2x9kl"] Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.742972 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.745433 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.760959 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568675b579-2x9kl"] Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.895127 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df4kz\" (UniqueName: \"kubernetes.io/projected/693932a0-7177-4647-9581-6f56e9446b95-kube-api-access-df4kz\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.895186 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-nb\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.895228 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-config\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.895252 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-openstack-edpm-ipam\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.895270 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-sb\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.895310 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-dns-svc\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.997037 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df4kz\" (UniqueName: \"kubernetes.io/projected/693932a0-7177-4647-9581-6f56e9446b95-kube-api-access-df4kz\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.997119 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-nb\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.997165 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-config\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.997184 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-openstack-edpm-ipam\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.997202 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-sb\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.997240 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-dns-svc\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.998344 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-openstack-edpm-ipam\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.998480 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-dns-svc\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.998539 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-nb\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.998981 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-config\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:55 crc kubenswrapper[4726]: I1123 20:28:55.998997 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-sb\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:56 crc kubenswrapper[4726]: I1123 20:28:56.021735 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df4kz\" (UniqueName: \"kubernetes.io/projected/693932a0-7177-4647-9581-6f56e9446b95-kube-api-access-df4kz\") pod \"dnsmasq-dns-568675b579-2x9kl\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:56 crc kubenswrapper[4726]: I1123 20:28:56.095740 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:56 crc kubenswrapper[4726]: I1123 20:28:56.325760 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0883d1c4-a52b-4221-a651-8ba8c5463dd6","Type":"ContainerStarted","Data":"1594d44f3a69bc835662a5337b13e5bf40fd879e7af9cebdba7f992a5bfa3b84"} Nov 23 20:28:56 crc kubenswrapper[4726]: I1123 20:28:56.603627 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568675b579-2x9kl"] Nov 23 20:28:57 crc kubenswrapper[4726]: I1123 20:28:57.343225 4726 generic.go:334] "Generic (PLEG): container finished" podID="693932a0-7177-4647-9581-6f56e9446b95" containerID="20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd" exitCode=0 Nov 23 20:28:57 crc kubenswrapper[4726]: I1123 20:28:57.343311 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-2x9kl" event={"ID":"693932a0-7177-4647-9581-6f56e9446b95","Type":"ContainerDied","Data":"20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd"} Nov 23 20:28:57 crc kubenswrapper[4726]: I1123 20:28:57.343618 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-2x9kl" event={"ID":"693932a0-7177-4647-9581-6f56e9446b95","Type":"ContainerStarted","Data":"3fbee78470abee74c489c0f6b1313d2bfe71a8ec247b8b60941069f1f84ae074"} Nov 23 20:28:58 crc kubenswrapper[4726]: I1123 20:28:58.380553 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-2x9kl" event={"ID":"693932a0-7177-4647-9581-6f56e9446b95","Type":"ContainerStarted","Data":"788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0"} Nov 23 20:28:58 crc kubenswrapper[4726]: I1123 20:28:58.381541 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:28:58 crc kubenswrapper[4726]: I1123 20:28:58.427989 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568675b579-2x9kl" podStartSLOduration=3.427946953 podStartE2EDuration="3.427946953s" podCreationTimestamp="2025-11-23 20:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:28:58.419650484 +0000 UTC m=+1246.568691510" watchObservedRunningTime="2025-11-23 20:28:58.427946953 +0000 UTC m=+1246.576987979" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.097131 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.201992 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-bzgd2"] Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.366863 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5db79bb959-dxv4q"] Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.368619 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.379398 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5db79bb959-dxv4q"] Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.472784 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" podUID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" containerName="dnsmasq-dns" containerID="cri-o://b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771" gracePeriod=10 Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.516629 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-dns-svc\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.516686 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt4v5\" (UniqueName: \"kubernetes.io/projected/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-kube-api-access-bt4v5\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.516812 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-sb\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.516913 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-nb\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.516944 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-config\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.516965 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-openstack-edpm-ipam\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.618651 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-dns-svc\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.618729 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt4v5\" (UniqueName: \"kubernetes.io/projected/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-kube-api-access-bt4v5\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.618770 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-sb\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.618795 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-nb\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.618819 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-config\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.618839 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-openstack-edpm-ipam\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.619698 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-dns-svc\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.619701 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-openstack-edpm-ipam\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.619990 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-sb\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.620169 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-nb\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.621208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-config\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.640559 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt4v5\" (UniqueName: \"kubernetes.io/projected/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-kube-api-access-bt4v5\") pod \"dnsmasq-dns-5db79bb959-dxv4q\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.691549 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:06 crc kubenswrapper[4726]: I1123 20:29:06.887809 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.024664 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqnmj\" (UniqueName: \"kubernetes.io/projected/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-kube-api-access-mqnmj\") pod \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.024789 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-nb\") pod \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.025397 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-sb\") pod \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.025452 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-config\") pod \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.025525 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-dns-svc\") pod \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\" (UID: \"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a\") " Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.029699 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-kube-api-access-mqnmj" (OuterVolumeSpecName: "kube-api-access-mqnmj") pod "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" (UID: "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a"). InnerVolumeSpecName "kube-api-access-mqnmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.071330 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" (UID: "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.071371 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-config" (OuterVolumeSpecName: "config") pod "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" (UID: "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.078399 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" (UID: "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.079617 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" (UID: "78b61c97-0e49-4b9c-abd7-8e4963a9fd2a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.127162 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.127210 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqnmj\" (UniqueName: \"kubernetes.io/projected/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-kube-api-access-mqnmj\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.127226 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.127237 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.127248 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.293756 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5db79bb959-dxv4q"] Nov 23 20:29:07 crc kubenswrapper[4726]: W1123 20:29:07.300375 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9970b11b_0b4c_49b6_868e_7ee0ad82a9ef.slice/crio-2c4e4d8ce47b445a8979af7d2537bf21cf5ac88e0223ccd60efca8483df9547c WatchSource:0}: Error finding container 2c4e4d8ce47b445a8979af7d2537bf21cf5ac88e0223ccd60efca8483df9547c: Status 404 returned error can't find the container with id 2c4e4d8ce47b445a8979af7d2537bf21cf5ac88e0223ccd60efca8483df9547c Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.484727 4726 generic.go:334] "Generic (PLEG): container finished" podID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" containerID="b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771" exitCode=0 Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.484781 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" event={"ID":"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a","Type":"ContainerDied","Data":"b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771"} Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.484829 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" event={"ID":"78b61c97-0e49-4b9c-abd7-8e4963a9fd2a","Type":"ContainerDied","Data":"d73f428227a5f562f2b18cecc50b6363753595e1a1707fcb13c742acfe5ab37c"} Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.484827 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c9b558957-bzgd2" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.484893 4726 scope.go:117] "RemoveContainer" containerID="b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.488240 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" event={"ID":"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef","Type":"ContainerStarted","Data":"d916aca1be7cb270e111750a6edee56ec2ed05ff816e7f944f99b4e8c94e53e9"} Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.488285 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" event={"ID":"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef","Type":"ContainerStarted","Data":"2c4e4d8ce47b445a8979af7d2537bf21cf5ac88e0223ccd60efca8483df9547c"} Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.534838 4726 scope.go:117] "RemoveContainer" containerID="efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.555382 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-bzgd2"] Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.567333 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c9b558957-bzgd2"] Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.571458 4726 scope.go:117] "RemoveContainer" containerID="b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771" Nov 23 20:29:07 crc kubenswrapper[4726]: E1123 20:29:07.571996 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771\": container with ID starting with b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771 not found: ID does not exist" containerID="b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.572047 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771"} err="failed to get container status \"b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771\": rpc error: code = NotFound desc = could not find container \"b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771\": container with ID starting with b4eecec2fda595c301a462a83c1ee3d50fe4cbf32aff8a1c80c1e087297da771 not found: ID does not exist" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.572072 4726 scope.go:117] "RemoveContainer" containerID="efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9" Nov 23 20:29:07 crc kubenswrapper[4726]: E1123 20:29:07.572525 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9\": container with ID starting with efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9 not found: ID does not exist" containerID="efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9" Nov 23 20:29:07 crc kubenswrapper[4726]: I1123 20:29:07.572561 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9"} err="failed to get container status \"efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9\": rpc error: code = NotFound desc = could not find container \"efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9\": container with ID starting with efdbe3a1e91f439b5bfe26a24ff0b7086d1e8f9d1e5cd010ea47b4e3563c2cf9 not found: ID does not exist" Nov 23 20:29:08 crc kubenswrapper[4726]: I1123 20:29:08.502073 4726 generic.go:334] "Generic (PLEG): container finished" podID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" containerID="d916aca1be7cb270e111750a6edee56ec2ed05ff816e7f944f99b4e8c94e53e9" exitCode=0 Nov 23 20:29:08 crc kubenswrapper[4726]: I1123 20:29:08.502133 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" event={"ID":"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef","Type":"ContainerDied","Data":"d916aca1be7cb270e111750a6edee56ec2ed05ff816e7f944f99b4e8c94e53e9"} Nov 23 20:29:08 crc kubenswrapper[4726]: I1123 20:29:08.606885 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" path="/var/lib/kubelet/pods/78b61c97-0e49-4b9c-abd7-8e4963a9fd2a/volumes" Nov 23 20:29:09 crc kubenswrapper[4726]: I1123 20:29:09.515354 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" event={"ID":"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef","Type":"ContainerStarted","Data":"0c54114e02a79af89484aae3ce55a2532c71913989ade53e32ce3596a125d907"} Nov 23 20:29:09 crc kubenswrapper[4726]: I1123 20:29:09.515907 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:09 crc kubenswrapper[4726]: I1123 20:29:09.554821 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" podStartSLOduration=3.554796752 podStartE2EDuration="3.554796752s" podCreationTimestamp="2025-11-23 20:29:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:29:09.540144317 +0000 UTC m=+1257.689185283" watchObservedRunningTime="2025-11-23 20:29:09.554796752 +0000 UTC m=+1257.703837708" Nov 23 20:29:16 crc kubenswrapper[4726]: I1123 20:29:16.693911 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 20:29:16 crc kubenswrapper[4726]: I1123 20:29:16.779832 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568675b579-2x9kl"] Nov 23 20:29:16 crc kubenswrapper[4726]: I1123 20:29:16.780749 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568675b579-2x9kl" podUID="693932a0-7177-4647-9581-6f56e9446b95" containerName="dnsmasq-dns" containerID="cri-o://788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0" gracePeriod=10 Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.272285 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.354452 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-dns-svc\") pod \"693932a0-7177-4647-9581-6f56e9446b95\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.354822 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-config\") pod \"693932a0-7177-4647-9581-6f56e9446b95\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.354897 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-openstack-edpm-ipam\") pod \"693932a0-7177-4647-9581-6f56e9446b95\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.354927 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-sb\") pod \"693932a0-7177-4647-9581-6f56e9446b95\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.354964 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-nb\") pod \"693932a0-7177-4647-9581-6f56e9446b95\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.355010 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df4kz\" (UniqueName: \"kubernetes.io/projected/693932a0-7177-4647-9581-6f56e9446b95-kube-api-access-df4kz\") pod \"693932a0-7177-4647-9581-6f56e9446b95\" (UID: \"693932a0-7177-4647-9581-6f56e9446b95\") " Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.368704 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/693932a0-7177-4647-9581-6f56e9446b95-kube-api-access-df4kz" (OuterVolumeSpecName: "kube-api-access-df4kz") pod "693932a0-7177-4647-9581-6f56e9446b95" (UID: "693932a0-7177-4647-9581-6f56e9446b95"). InnerVolumeSpecName "kube-api-access-df4kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.408596 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "693932a0-7177-4647-9581-6f56e9446b95" (UID: "693932a0-7177-4647-9581-6f56e9446b95"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.413980 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-config" (OuterVolumeSpecName: "config") pod "693932a0-7177-4647-9581-6f56e9446b95" (UID: "693932a0-7177-4647-9581-6f56e9446b95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.421482 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "693932a0-7177-4647-9581-6f56e9446b95" (UID: "693932a0-7177-4647-9581-6f56e9446b95"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.422124 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "693932a0-7177-4647-9581-6f56e9446b95" (UID: "693932a0-7177-4647-9581-6f56e9446b95"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.423603 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "693932a0-7177-4647-9581-6f56e9446b95" (UID: "693932a0-7177-4647-9581-6f56e9446b95"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.457194 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-config\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.457224 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.457236 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.457247 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.457256 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df4kz\" (UniqueName: \"kubernetes.io/projected/693932a0-7177-4647-9581-6f56e9446b95-kube-api-access-df4kz\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.457265 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/693932a0-7177-4647-9581-6f56e9446b95-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.607684 4726 generic.go:334] "Generic (PLEG): container finished" podID="693932a0-7177-4647-9581-6f56e9446b95" containerID="788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0" exitCode=0 Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.607749 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-2x9kl" event={"ID":"693932a0-7177-4647-9581-6f56e9446b95","Type":"ContainerDied","Data":"788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0"} Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.607784 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568675b579-2x9kl" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.607829 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568675b579-2x9kl" event={"ID":"693932a0-7177-4647-9581-6f56e9446b95","Type":"ContainerDied","Data":"3fbee78470abee74c489c0f6b1313d2bfe71a8ec247b8b60941069f1f84ae074"} Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.607907 4726 scope.go:117] "RemoveContainer" containerID="788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.639046 4726 scope.go:117] "RemoveContainer" containerID="20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.649262 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568675b579-2x9kl"] Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.662614 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568675b579-2x9kl"] Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.683293 4726 scope.go:117] "RemoveContainer" containerID="788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0" Nov 23 20:29:17 crc kubenswrapper[4726]: E1123 20:29:17.683763 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0\": container with ID starting with 788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0 not found: ID does not exist" containerID="788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.683807 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0"} err="failed to get container status \"788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0\": rpc error: code = NotFound desc = could not find container \"788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0\": container with ID starting with 788155b7eb6b9db4cd750169cbeb5209c3533447014d9f50dc1c2d20b2bbe9e0 not found: ID does not exist" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.683831 4726 scope.go:117] "RemoveContainer" containerID="20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd" Nov 23 20:29:17 crc kubenswrapper[4726]: E1123 20:29:17.684257 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd\": container with ID starting with 20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd not found: ID does not exist" containerID="20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd" Nov 23 20:29:17 crc kubenswrapper[4726]: I1123 20:29:17.684308 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd"} err="failed to get container status \"20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd\": rpc error: code = NotFound desc = could not find container \"20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd\": container with ID starting with 20f9ec89ec1db4591e83b765bfd641a0ae4c9bc581ab7dcb81e2606a8cc47cbd not found: ID does not exist" Nov 23 20:29:18 crc kubenswrapper[4726]: I1123 20:29:18.598538 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="693932a0-7177-4647-9581-6f56e9446b95" path="/var/lib/kubelet/pods/693932a0-7177-4647-9581-6f56e9446b95/volumes" Nov 23 20:29:26 crc kubenswrapper[4726]: I1123 20:29:26.724781 4726 generic.go:334] "Generic (PLEG): container finished" podID="f962af44-2b42-4924-81b8-872c96335906" containerID="908c9a60dbbc81790112cf7531d732f69abb86cee42b37a5f3087b22aea3b991" exitCode=0 Nov 23 20:29:26 crc kubenswrapper[4726]: I1123 20:29:26.724899 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f962af44-2b42-4924-81b8-872c96335906","Type":"ContainerDied","Data":"908c9a60dbbc81790112cf7531d732f69abb86cee42b37a5f3087b22aea3b991"} Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.032428 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk"] Nov 23 20:29:27 crc kubenswrapper[4726]: E1123 20:29:27.033200 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" containerName="dnsmasq-dns" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.033220 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" containerName="dnsmasq-dns" Nov 23 20:29:27 crc kubenswrapper[4726]: E1123 20:29:27.033234 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693932a0-7177-4647-9581-6f56e9446b95" containerName="init" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.033241 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="693932a0-7177-4647-9581-6f56e9446b95" containerName="init" Nov 23 20:29:27 crc kubenswrapper[4726]: E1123 20:29:27.033257 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" containerName="init" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.033264 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" containerName="init" Nov 23 20:29:27 crc kubenswrapper[4726]: E1123 20:29:27.033278 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693932a0-7177-4647-9581-6f56e9446b95" containerName="dnsmasq-dns" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.033284 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="693932a0-7177-4647-9581-6f56e9446b95" containerName="dnsmasq-dns" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.033483 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="78b61c97-0e49-4b9c-abd7-8e4963a9fd2a" containerName="dnsmasq-dns" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.033509 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="693932a0-7177-4647-9581-6f56e9446b95" containerName="dnsmasq-dns" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.034243 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.038029 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.038241 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.038398 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.044520 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.047207 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk"] Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.184911 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6vzd\" (UniqueName: \"kubernetes.io/projected/00328a0d-972a-4fcf-9889-58fbac5548b7-kube-api-access-r6vzd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.185060 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.185242 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.185353 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.287789 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6vzd\" (UniqueName: \"kubernetes.io/projected/00328a0d-972a-4fcf-9889-58fbac5548b7-kube-api-access-r6vzd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.287890 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.287962 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.288018 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.294570 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.294968 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.295366 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.320713 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6vzd\" (UniqueName: \"kubernetes.io/projected/00328a0d-972a-4fcf-9889-58fbac5548b7-kube-api-access-r6vzd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.353248 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.737010 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f962af44-2b42-4924-81b8-872c96335906","Type":"ContainerStarted","Data":"5c797070e1f7daac27c26f19e2b672d117cdb18957c8af82d6b1e2310a78e315"} Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.737858 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.763090 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.763072426 podStartE2EDuration="36.763072426s" podCreationTimestamp="2025-11-23 20:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:29:27.762581042 +0000 UTC m=+1275.911622018" watchObservedRunningTime="2025-11-23 20:29:27.763072426 +0000 UTC m=+1275.912113382" Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.949442 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk"] Nov 23 20:29:27 crc kubenswrapper[4726]: I1123 20:29:27.962999 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 20:29:28 crc kubenswrapper[4726]: I1123 20:29:28.753537 4726 generic.go:334] "Generic (PLEG): container finished" podID="0883d1c4-a52b-4221-a651-8ba8c5463dd6" containerID="1594d44f3a69bc835662a5337b13e5bf40fd879e7af9cebdba7f992a5bfa3b84" exitCode=0 Nov 23 20:29:28 crc kubenswrapper[4726]: I1123 20:29:28.754032 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0883d1c4-a52b-4221-a651-8ba8c5463dd6","Type":"ContainerDied","Data":"1594d44f3a69bc835662a5337b13e5bf40fd879e7af9cebdba7f992a5bfa3b84"} Nov 23 20:29:28 crc kubenswrapper[4726]: I1123 20:29:28.762729 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" event={"ID":"00328a0d-972a-4fcf-9889-58fbac5548b7","Type":"ContainerStarted","Data":"2ca4dee8260e943dfa52bacab22a1604e07615a12343f612a987be8d27aac036"} Nov 23 20:29:29 crc kubenswrapper[4726]: I1123 20:29:29.773425 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0883d1c4-a52b-4221-a651-8ba8c5463dd6","Type":"ContainerStarted","Data":"a02f147636c9f3063b5aa743e79147297179f534e8d1871cd02539cbef92036e"} Nov 23 20:29:29 crc kubenswrapper[4726]: I1123 20:29:29.775181 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:29:29 crc kubenswrapper[4726]: I1123 20:29:29.811132 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.811108452 podStartE2EDuration="36.811108452s" podCreationTimestamp="2025-11-23 20:28:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:29:29.800545 +0000 UTC m=+1277.949585956" watchObservedRunningTime="2025-11-23 20:29:29.811108452 +0000 UTC m=+1277.960149408" Nov 23 20:29:39 crc kubenswrapper[4726]: I1123 20:29:39.044272 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:29:39 crc kubenswrapper[4726]: I1123 20:29:39.044779 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:29:40 crc kubenswrapper[4726]: I1123 20:29:40.911911 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" event={"ID":"00328a0d-972a-4fcf-9889-58fbac5548b7","Type":"ContainerStarted","Data":"c7c2efa2d1aa7d18e62b6de42d411a1e5e1c671d1271cb47d13b2aba9c043cb7"} Nov 23 20:29:40 crc kubenswrapper[4726]: I1123 20:29:40.935424 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" podStartSLOduration=3.138887583 podStartE2EDuration="14.935403739s" podCreationTimestamp="2025-11-23 20:29:26 +0000 UTC" firstStartedPulling="2025-11-23 20:29:27.962550165 +0000 UTC m=+1276.111591121" lastFinishedPulling="2025-11-23 20:29:39.759066311 +0000 UTC m=+1287.908107277" observedRunningTime="2025-11-23 20:29:40.92926475 +0000 UTC m=+1289.078305716" watchObservedRunningTime="2025-11-23 20:29:40.935403739 +0000 UTC m=+1289.084444705" Nov 23 20:29:41 crc kubenswrapper[4726]: I1123 20:29:41.656177 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 23 20:29:43 crc kubenswrapper[4726]: I1123 20:29:43.782220 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 23 20:29:52 crc kubenswrapper[4726]: I1123 20:29:52.079937 4726 generic.go:334] "Generic (PLEG): container finished" podID="00328a0d-972a-4fcf-9889-58fbac5548b7" containerID="c7c2efa2d1aa7d18e62b6de42d411a1e5e1c671d1271cb47d13b2aba9c043cb7" exitCode=0 Nov 23 20:29:52 crc kubenswrapper[4726]: I1123 20:29:52.080484 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" event={"ID":"00328a0d-972a-4fcf-9889-58fbac5548b7","Type":"ContainerDied","Data":"c7c2efa2d1aa7d18e62b6de42d411a1e5e1c671d1271cb47d13b2aba9c043cb7"} Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.506651 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.586119 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-ssh-key\") pod \"00328a0d-972a-4fcf-9889-58fbac5548b7\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.586229 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-repo-setup-combined-ca-bundle\") pod \"00328a0d-972a-4fcf-9889-58fbac5548b7\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.586281 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-inventory\") pod \"00328a0d-972a-4fcf-9889-58fbac5548b7\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.586298 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6vzd\" (UniqueName: \"kubernetes.io/projected/00328a0d-972a-4fcf-9889-58fbac5548b7-kube-api-access-r6vzd\") pod \"00328a0d-972a-4fcf-9889-58fbac5548b7\" (UID: \"00328a0d-972a-4fcf-9889-58fbac5548b7\") " Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.594695 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00328a0d-972a-4fcf-9889-58fbac5548b7-kube-api-access-r6vzd" (OuterVolumeSpecName: "kube-api-access-r6vzd") pod "00328a0d-972a-4fcf-9889-58fbac5548b7" (UID: "00328a0d-972a-4fcf-9889-58fbac5548b7"). InnerVolumeSpecName "kube-api-access-r6vzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.603044 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "00328a0d-972a-4fcf-9889-58fbac5548b7" (UID: "00328a0d-972a-4fcf-9889-58fbac5548b7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.624372 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "00328a0d-972a-4fcf-9889-58fbac5548b7" (UID: "00328a0d-972a-4fcf-9889-58fbac5548b7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.643540 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-inventory" (OuterVolumeSpecName: "inventory") pod "00328a0d-972a-4fcf-9889-58fbac5548b7" (UID: "00328a0d-972a-4fcf-9889-58fbac5548b7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.688789 4726 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.688825 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.688837 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6vzd\" (UniqueName: \"kubernetes.io/projected/00328a0d-972a-4fcf-9889-58fbac5548b7-kube-api-access-r6vzd\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:53 crc kubenswrapper[4726]: I1123 20:29:53.688847 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00328a0d-972a-4fcf-9889-58fbac5548b7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.103501 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" event={"ID":"00328a0d-972a-4fcf-9889-58fbac5548b7","Type":"ContainerDied","Data":"2ca4dee8260e943dfa52bacab22a1604e07615a12343f612a987be8d27aac036"} Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.103550 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ca4dee8260e943dfa52bacab22a1604e07615a12343f612a987be8d27aac036" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.103560 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.170150 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc"] Nov 23 20:29:54 crc kubenswrapper[4726]: E1123 20:29:54.170513 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00328a0d-972a-4fcf-9889-58fbac5548b7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.170533 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="00328a0d-972a-4fcf-9889-58fbac5548b7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.170742 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="00328a0d-972a-4fcf-9889-58fbac5548b7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.171366 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.173161 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.173772 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.174321 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.180071 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.191535 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc"] Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.321506 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.321554 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcxzp\" (UniqueName: \"kubernetes.io/projected/0ff7e920-cb8b-453e-82cb-454e59325a43-kube-api-access-bcxzp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.321617 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.321645 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.423100 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.423149 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcxzp\" (UniqueName: \"kubernetes.io/projected/0ff7e920-cb8b-453e-82cb-454e59325a43-kube-api-access-bcxzp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.423195 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.423218 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.427276 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.427355 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.433679 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.442443 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcxzp\" (UniqueName: \"kubernetes.io/projected/0ff7e920-cb8b-453e-82cb-454e59325a43-kube-api-access-bcxzp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.489939 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:29:54 crc kubenswrapper[4726]: I1123 20:29:54.816308 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc"] Nov 23 20:29:55 crc kubenswrapper[4726]: I1123 20:29:55.112780 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" event={"ID":"0ff7e920-cb8b-453e-82cb-454e59325a43","Type":"ContainerStarted","Data":"925fcb30dec6c94fc33872d2d21a7522fa7fe1dacb1dbb373b44c99d1338b0ce"} Nov 23 20:29:56 crc kubenswrapper[4726]: I1123 20:29:56.124713 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" event={"ID":"0ff7e920-cb8b-453e-82cb-454e59325a43","Type":"ContainerStarted","Data":"740fcead03fe7884ee89023a0f11ad53bdc4e850757cd8a8a92051626d94f729"} Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.162511 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" podStartSLOduration=5.622313893 podStartE2EDuration="6.162473924s" podCreationTimestamp="2025-11-23 20:29:54 +0000 UTC" firstStartedPulling="2025-11-23 20:29:54.813588426 +0000 UTC m=+1302.962629382" lastFinishedPulling="2025-11-23 20:29:55.353748457 +0000 UTC m=+1303.502789413" observedRunningTime="2025-11-23 20:29:56.143412204 +0000 UTC m=+1304.292453200" watchObservedRunningTime="2025-11-23 20:30:00.162473924 +0000 UTC m=+1308.311514910" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.166644 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj"] Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.168856 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.171262 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.172417 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.176130 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b082b91-898d-411b-beb4-7c6482f4c169-config-volume\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.176249 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqfzb\" (UniqueName: \"kubernetes.io/projected/4b082b91-898d-411b-beb4-7c6482f4c169-kube-api-access-hqfzb\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.176366 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b082b91-898d-411b-beb4-7c6482f4c169-secret-volume\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.182495 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj"] Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.278046 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b082b91-898d-411b-beb4-7c6482f4c169-secret-volume\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.278148 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b082b91-898d-411b-beb4-7c6482f4c169-config-volume\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.278206 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqfzb\" (UniqueName: \"kubernetes.io/projected/4b082b91-898d-411b-beb4-7c6482f4c169-kube-api-access-hqfzb\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.281147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b082b91-898d-411b-beb4-7c6482f4c169-config-volume\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.288086 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b082b91-898d-411b-beb4-7c6482f4c169-secret-volume\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.295941 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqfzb\" (UniqueName: \"kubernetes.io/projected/4b082b91-898d-411b-beb4-7c6482f4c169-kube-api-access-hqfzb\") pod \"collect-profiles-29398830-6pqgj\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.533959 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:00 crc kubenswrapper[4726]: I1123 20:30:00.980194 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj"] Nov 23 20:30:01 crc kubenswrapper[4726]: I1123 20:30:01.177822 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" event={"ID":"4b082b91-898d-411b-beb4-7c6482f4c169","Type":"ContainerStarted","Data":"ef63d815221c7aa0e2bab66001a7ebd5bc73bbc66c7bdb599a0234fff6037a2a"} Nov 23 20:30:01 crc kubenswrapper[4726]: I1123 20:30:01.177897 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" event={"ID":"4b082b91-898d-411b-beb4-7c6482f4c169","Type":"ContainerStarted","Data":"e7879f080b159e4560a5d1e639c13b944ed9440d49a3295c4816b5632c15ddad"} Nov 23 20:30:01 crc kubenswrapper[4726]: I1123 20:30:01.200095 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" podStartSLOduration=1.200072633 podStartE2EDuration="1.200072633s" podCreationTimestamp="2025-11-23 20:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 20:30:01.192056359 +0000 UTC m=+1309.341097325" watchObservedRunningTime="2025-11-23 20:30:01.200072633 +0000 UTC m=+1309.349113589" Nov 23 20:30:02 crc kubenswrapper[4726]: I1123 20:30:02.201616 4726 generic.go:334] "Generic (PLEG): container finished" podID="4b082b91-898d-411b-beb4-7c6482f4c169" containerID="ef63d815221c7aa0e2bab66001a7ebd5bc73bbc66c7bdb599a0234fff6037a2a" exitCode=0 Nov 23 20:30:02 crc kubenswrapper[4726]: I1123 20:30:02.201703 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" event={"ID":"4b082b91-898d-411b-beb4-7c6482f4c169","Type":"ContainerDied","Data":"ef63d815221c7aa0e2bab66001a7ebd5bc73bbc66c7bdb599a0234fff6037a2a"} Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.557412 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.736272 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b082b91-898d-411b-beb4-7c6482f4c169-secret-volume\") pod \"4b082b91-898d-411b-beb4-7c6482f4c169\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.736599 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b082b91-898d-411b-beb4-7c6482f4c169-config-volume\") pod \"4b082b91-898d-411b-beb4-7c6482f4c169\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.736752 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqfzb\" (UniqueName: \"kubernetes.io/projected/4b082b91-898d-411b-beb4-7c6482f4c169-kube-api-access-hqfzb\") pod \"4b082b91-898d-411b-beb4-7c6482f4c169\" (UID: \"4b082b91-898d-411b-beb4-7c6482f4c169\") " Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.737088 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b082b91-898d-411b-beb4-7c6482f4c169-config-volume" (OuterVolumeSpecName: "config-volume") pod "4b082b91-898d-411b-beb4-7c6482f4c169" (UID: "4b082b91-898d-411b-beb4-7c6482f4c169"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.745795 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b082b91-898d-411b-beb4-7c6482f4c169-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4b082b91-898d-411b-beb4-7c6482f4c169" (UID: "4b082b91-898d-411b-beb4-7c6482f4c169"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.747352 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b082b91-898d-411b-beb4-7c6482f4c169-kube-api-access-hqfzb" (OuterVolumeSpecName: "kube-api-access-hqfzb") pod "4b082b91-898d-411b-beb4-7c6482f4c169" (UID: "4b082b91-898d-411b-beb4-7c6482f4c169"). InnerVolumeSpecName "kube-api-access-hqfzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.838660 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b082b91-898d-411b-beb4-7c6482f4c169-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.838700 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b082b91-898d-411b-beb4-7c6482f4c169-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:30:03 crc kubenswrapper[4726]: I1123 20:30:03.838714 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqfzb\" (UniqueName: \"kubernetes.io/projected/4b082b91-898d-411b-beb4-7c6482f4c169-kube-api-access-hqfzb\") on node \"crc\" DevicePath \"\"" Nov 23 20:30:04 crc kubenswrapper[4726]: I1123 20:30:04.234837 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" event={"ID":"4b082b91-898d-411b-beb4-7c6482f4c169","Type":"ContainerDied","Data":"e7879f080b159e4560a5d1e639c13b944ed9440d49a3295c4816b5632c15ddad"} Nov 23 20:30:04 crc kubenswrapper[4726]: I1123 20:30:04.234918 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7879f080b159e4560a5d1e639c13b944ed9440d49a3295c4816b5632c15ddad" Nov 23 20:30:04 crc kubenswrapper[4726]: I1123 20:30:04.234977 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj" Nov 23 20:30:09 crc kubenswrapper[4726]: I1123 20:30:09.042735 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:30:09 crc kubenswrapper[4726]: I1123 20:30:09.043411 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.042614 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.043291 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.043353 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.044205 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"828799a4d998638471568a30864015defe9240aef8698bc3b256f460fe153d23"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.044278 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://828799a4d998638471568a30864015defe9240aef8698bc3b256f460fe153d23" gracePeriod=600 Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.677021 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="828799a4d998638471568a30864015defe9240aef8698bc3b256f460fe153d23" exitCode=0 Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.677074 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"828799a4d998638471568a30864015defe9240aef8698bc3b256f460fe153d23"} Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.677365 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa"} Nov 23 20:30:39 crc kubenswrapper[4726]: I1123 20:30:39.677430 4726 scope.go:117] "RemoveContainer" containerID="93d2133e107e131190f68c969bc423a889dae2dbf2757948997b4402e77bc885" Nov 23 20:31:16 crc kubenswrapper[4726]: I1123 20:31:16.664429 4726 scope.go:117] "RemoveContainer" containerID="e728c2dae0ea69563bc7191687530716dfa0097e064bf8a3f753e86f503db4f7" Nov 23 20:31:16 crc kubenswrapper[4726]: I1123 20:31:16.700946 4726 scope.go:117] "RemoveContainer" containerID="db4e8ff916ad6ec86fe213933be0253ba46268725cce9a2601df99d2f387c4fa" Nov 23 20:31:16 crc kubenswrapper[4726]: I1123 20:31:16.777653 4726 scope.go:117] "RemoveContainer" containerID="ea52e905d86a603a499077a3ba92ccddae806f3d27b4e25ca709a5dbda1af821" Nov 23 20:31:16 crc kubenswrapper[4726]: I1123 20:31:16.827465 4726 scope.go:117] "RemoveContainer" containerID="27d51503001e6dc16b2018d4a060cc21f99e42ec9d6537683ff985b7654e9697" Nov 23 20:31:16 crc kubenswrapper[4726]: I1123 20:31:16.859228 4726 scope.go:117] "RemoveContainer" containerID="e658d0749861ccc937b6325f5382fbd538c49a561a8aae44ab581690dd14187b" Nov 23 20:31:16 crc kubenswrapper[4726]: I1123 20:31:16.885541 4726 scope.go:117] "RemoveContainer" containerID="aac7e9a0984302acfad53201f5828de638a7de333ad6b6d93667b6b7a76fcad3" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.280249 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lgmvk"] Nov 23 20:31:50 crc kubenswrapper[4726]: E1123 20:31:50.287668 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b082b91-898d-411b-beb4-7c6482f4c169" containerName="collect-profiles" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.287997 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b082b91-898d-411b-beb4-7c6482f4c169" containerName="collect-profiles" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.288502 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b082b91-898d-411b-beb4-7c6482f4c169" containerName="collect-profiles" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.292065 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.296360 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgmvk"] Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.483043 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-utilities\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.483314 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmnjh\" (UniqueName: \"kubernetes.io/projected/37272200-e66f-4b21-b337-ff745508d63e-kube-api-access-fmnjh\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.483359 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-catalog-content\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.584465 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-catalog-content\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.584633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-utilities\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.584690 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmnjh\" (UniqueName: \"kubernetes.io/projected/37272200-e66f-4b21-b337-ff745508d63e-kube-api-access-fmnjh\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.585220 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-catalog-content\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.585224 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-utilities\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.614630 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmnjh\" (UniqueName: \"kubernetes.io/projected/37272200-e66f-4b21-b337-ff745508d63e-kube-api-access-fmnjh\") pod \"redhat-operators-lgmvk\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:50 crc kubenswrapper[4726]: I1123 20:31:50.631630 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:31:51 crc kubenswrapper[4726]: I1123 20:31:51.125153 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgmvk"] Nov 23 20:31:51 crc kubenswrapper[4726]: I1123 20:31:51.455748 4726 generic.go:334] "Generic (PLEG): container finished" podID="37272200-e66f-4b21-b337-ff745508d63e" containerID="a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9" exitCode=0 Nov 23 20:31:51 crc kubenswrapper[4726]: I1123 20:31:51.455811 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgmvk" event={"ID":"37272200-e66f-4b21-b337-ff745508d63e","Type":"ContainerDied","Data":"a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9"} Nov 23 20:31:51 crc kubenswrapper[4726]: I1123 20:31:51.456089 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgmvk" event={"ID":"37272200-e66f-4b21-b337-ff745508d63e","Type":"ContainerStarted","Data":"6ba61b9e3759e875c3e55e75d748686c7711b2daa58782b985bb35ce01a1fdf6"} Nov 23 20:31:52 crc kubenswrapper[4726]: I1123 20:31:52.465768 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgmvk" event={"ID":"37272200-e66f-4b21-b337-ff745508d63e","Type":"ContainerStarted","Data":"3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b"} Nov 23 20:31:56 crc kubenswrapper[4726]: I1123 20:31:56.525220 4726 generic.go:334] "Generic (PLEG): container finished" podID="37272200-e66f-4b21-b337-ff745508d63e" containerID="3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b" exitCode=0 Nov 23 20:31:56 crc kubenswrapper[4726]: I1123 20:31:56.525591 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgmvk" event={"ID":"37272200-e66f-4b21-b337-ff745508d63e","Type":"ContainerDied","Data":"3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b"} Nov 23 20:31:57 crc kubenswrapper[4726]: I1123 20:31:57.541126 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgmvk" event={"ID":"37272200-e66f-4b21-b337-ff745508d63e","Type":"ContainerStarted","Data":"87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea"} Nov 23 20:31:57 crc kubenswrapper[4726]: I1123 20:31:57.566078 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lgmvk" podStartSLOduration=2.118896412 podStartE2EDuration="7.566049203s" podCreationTimestamp="2025-11-23 20:31:50 +0000 UTC" firstStartedPulling="2025-11-23 20:31:51.457855795 +0000 UTC m=+1419.606896751" lastFinishedPulling="2025-11-23 20:31:56.905008586 +0000 UTC m=+1425.054049542" observedRunningTime="2025-11-23 20:31:57.561338211 +0000 UTC m=+1425.710379197" watchObservedRunningTime="2025-11-23 20:31:57.566049203 +0000 UTC m=+1425.715090199" Nov 23 20:32:00 crc kubenswrapper[4726]: I1123 20:32:00.632853 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:32:00 crc kubenswrapper[4726]: I1123 20:32:00.633126 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:32:01 crc kubenswrapper[4726]: I1123 20:32:01.707904 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lgmvk" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="registry-server" probeResult="failure" output=< Nov 23 20:32:01 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 20:32:01 crc kubenswrapper[4726]: > Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.700760 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6z5fs"] Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.704231 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.723787 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6z5fs"] Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.825598 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-utilities\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.826088 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxh88\" (UniqueName: \"kubernetes.io/projected/10f1a285-434b-4044-9f81-67058a1fc70c-kube-api-access-rxh88\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.826414 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-catalog-content\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.929107 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-utilities\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.929372 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxh88\" (UniqueName: \"kubernetes.io/projected/10f1a285-434b-4044-9f81-67058a1fc70c-kube-api-access-rxh88\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.929654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-catalog-content\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.929858 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-utilities\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.930305 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-catalog-content\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:08 crc kubenswrapper[4726]: I1123 20:32:08.968636 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxh88\" (UniqueName: \"kubernetes.io/projected/10f1a285-434b-4044-9f81-67058a1fc70c-kube-api-access-rxh88\") pod \"community-operators-6z5fs\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:09 crc kubenswrapper[4726]: I1123 20:32:09.023195 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:09 crc kubenswrapper[4726]: I1123 20:32:09.463836 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6z5fs"] Nov 23 20:32:09 crc kubenswrapper[4726]: I1123 20:32:09.671729 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5fs" event={"ID":"10f1a285-434b-4044-9f81-67058a1fc70c","Type":"ContainerStarted","Data":"cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8"} Nov 23 20:32:09 crc kubenswrapper[4726]: I1123 20:32:09.673043 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5fs" event={"ID":"10f1a285-434b-4044-9f81-67058a1fc70c","Type":"ContainerStarted","Data":"0898264973ace674f94b241ac7f6ee422574813af2affb3b1f602d8ead7ed9f8"} Nov 23 20:32:10 crc kubenswrapper[4726]: I1123 20:32:10.682240 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:32:10 crc kubenswrapper[4726]: I1123 20:32:10.689083 4726 generic.go:334] "Generic (PLEG): container finished" podID="10f1a285-434b-4044-9f81-67058a1fc70c" containerID="cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8" exitCode=0 Nov 23 20:32:10 crc kubenswrapper[4726]: I1123 20:32:10.690995 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5fs" event={"ID":"10f1a285-434b-4044-9f81-67058a1fc70c","Type":"ContainerDied","Data":"cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8"} Nov 23 20:32:10 crc kubenswrapper[4726]: I1123 20:32:10.691027 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5fs" event={"ID":"10f1a285-434b-4044-9f81-67058a1fc70c","Type":"ContainerStarted","Data":"e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d"} Nov 23 20:32:10 crc kubenswrapper[4726]: I1123 20:32:10.810619 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:32:12 crc kubenswrapper[4726]: I1123 20:32:12.714233 4726 generic.go:334] "Generic (PLEG): container finished" podID="10f1a285-434b-4044-9f81-67058a1fc70c" containerID="e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d" exitCode=0 Nov 23 20:32:12 crc kubenswrapper[4726]: I1123 20:32:12.714675 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5fs" event={"ID":"10f1a285-434b-4044-9f81-67058a1fc70c","Type":"ContainerDied","Data":"e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d"} Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.075269 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgmvk"] Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.075573 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lgmvk" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="registry-server" containerID="cri-o://87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea" gracePeriod=2 Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.537598 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.724457 4726 generic.go:334] "Generic (PLEG): container finished" podID="37272200-e66f-4b21-b337-ff745508d63e" containerID="87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea" exitCode=0 Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.724484 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-catalog-content\") pod \"37272200-e66f-4b21-b337-ff745508d63e\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.724850 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-utilities\") pod \"37272200-e66f-4b21-b337-ff745508d63e\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.724959 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmnjh\" (UniqueName: \"kubernetes.io/projected/37272200-e66f-4b21-b337-ff745508d63e-kube-api-access-fmnjh\") pod \"37272200-e66f-4b21-b337-ff745508d63e\" (UID: \"37272200-e66f-4b21-b337-ff745508d63e\") " Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.724636 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgmvk" event={"ID":"37272200-e66f-4b21-b337-ff745508d63e","Type":"ContainerDied","Data":"87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea"} Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.724662 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgmvk" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.726098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgmvk" event={"ID":"37272200-e66f-4b21-b337-ff745508d63e","Type":"ContainerDied","Data":"6ba61b9e3759e875c3e55e75d748686c7711b2daa58782b985bb35ce01a1fdf6"} Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.726145 4726 scope.go:117] "RemoveContainer" containerID="87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.726779 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-utilities" (OuterVolumeSpecName: "utilities") pod "37272200-e66f-4b21-b337-ff745508d63e" (UID: "37272200-e66f-4b21-b337-ff745508d63e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.733245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5fs" event={"ID":"10f1a285-434b-4044-9f81-67058a1fc70c","Type":"ContainerStarted","Data":"79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87"} Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.736690 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37272200-e66f-4b21-b337-ff745508d63e-kube-api-access-fmnjh" (OuterVolumeSpecName: "kube-api-access-fmnjh") pod "37272200-e66f-4b21-b337-ff745508d63e" (UID: "37272200-e66f-4b21-b337-ff745508d63e"). InnerVolumeSpecName "kube-api-access-fmnjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.791517 4726 scope.go:117] "RemoveContainer" containerID="3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.814164 4726 scope.go:117] "RemoveContainer" containerID="a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.825608 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37272200-e66f-4b21-b337-ff745508d63e" (UID: "37272200-e66f-4b21-b337-ff745508d63e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.827534 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.827563 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37272200-e66f-4b21-b337-ff745508d63e-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.827576 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmnjh\" (UniqueName: \"kubernetes.io/projected/37272200-e66f-4b21-b337-ff745508d63e-kube-api-access-fmnjh\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.854214 4726 scope.go:117] "RemoveContainer" containerID="87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea" Nov 23 20:32:13 crc kubenswrapper[4726]: E1123 20:32:13.854697 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea\": container with ID starting with 87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea not found: ID does not exist" containerID="87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.854812 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea"} err="failed to get container status \"87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea\": rpc error: code = NotFound desc = could not find container \"87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea\": container with ID starting with 87789ba2997b6156f6de8e3d9b9887fb92aee18fd590ea03cea4ebf20f7f90ea not found: ID does not exist" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.854939 4726 scope.go:117] "RemoveContainer" containerID="3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b" Nov 23 20:32:13 crc kubenswrapper[4726]: E1123 20:32:13.855359 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b\": container with ID starting with 3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b not found: ID does not exist" containerID="3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.855457 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b"} err="failed to get container status \"3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b\": rpc error: code = NotFound desc = could not find container \"3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b\": container with ID starting with 3348fbf474a3c4b8c9dc8ccfa12a9f7994a7e79a2e9ab0fe6448f6ec94ad6e9b not found: ID does not exist" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.855537 4726 scope.go:117] "RemoveContainer" containerID="a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9" Nov 23 20:32:13 crc kubenswrapper[4726]: E1123 20:32:13.855858 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9\": container with ID starting with a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9 not found: ID does not exist" containerID="a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9" Nov 23 20:32:13 crc kubenswrapper[4726]: I1123 20:32:13.855981 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9"} err="failed to get container status \"a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9\": rpc error: code = NotFound desc = could not find container \"a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9\": container with ID starting with a76a4580deb748140a96518af5e6851a3bdaf91b6fd4be7a566b219b9e7392f9 not found: ID does not exist" Nov 23 20:32:14 crc kubenswrapper[4726]: I1123 20:32:14.050849 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6z5fs" podStartSLOduration=2.590175007 podStartE2EDuration="6.050832115s" podCreationTimestamp="2025-11-23 20:32:08 +0000 UTC" firstStartedPulling="2025-11-23 20:32:09.677323625 +0000 UTC m=+1437.826364581" lastFinishedPulling="2025-11-23 20:32:13.137980733 +0000 UTC m=+1441.287021689" observedRunningTime="2025-11-23 20:32:13.759106614 +0000 UTC m=+1441.908147570" watchObservedRunningTime="2025-11-23 20:32:14.050832115 +0000 UTC m=+1442.199873071" Nov 23 20:32:14 crc kubenswrapper[4726]: I1123 20:32:14.057914 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgmvk"] Nov 23 20:32:14 crc kubenswrapper[4726]: I1123 20:32:14.064746 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lgmvk"] Nov 23 20:32:14 crc kubenswrapper[4726]: I1123 20:32:14.599280 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37272200-e66f-4b21-b337-ff745508d63e" path="/var/lib/kubelet/pods/37272200-e66f-4b21-b337-ff745508d63e/volumes" Nov 23 20:32:17 crc kubenswrapper[4726]: I1123 20:32:17.023558 4726 scope.go:117] "RemoveContainer" containerID="fdd4b982192768220a2598051eb750b7ba0396e2789c1d26e90fcc0f6b78069c" Nov 23 20:32:17 crc kubenswrapper[4726]: I1123 20:32:17.053521 4726 scope.go:117] "RemoveContainer" containerID="48f4363e2fa89c5a8408a6a7c21589d4c82008c2e40bf3f98d946ad9dc5b6304" Nov 23 20:32:17 crc kubenswrapper[4726]: I1123 20:32:17.103861 4726 scope.go:117] "RemoveContainer" containerID="77c17bb00fe8203cb0bded7db33549111b5f33c679b58deefd29ae1d8baa3c32" Nov 23 20:32:19 crc kubenswrapper[4726]: I1123 20:32:19.023394 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:19 crc kubenswrapper[4726]: I1123 20:32:19.023918 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:19 crc kubenswrapper[4726]: I1123 20:32:19.084826 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:19 crc kubenswrapper[4726]: I1123 20:32:19.834256 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:19 crc kubenswrapper[4726]: I1123 20:32:19.881256 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6z5fs"] Nov 23 20:32:21 crc kubenswrapper[4726]: I1123 20:32:21.827224 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6z5fs" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" containerName="registry-server" containerID="cri-o://79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87" gracePeriod=2 Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.407533 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.522858 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-catalog-content\") pod \"10f1a285-434b-4044-9f81-67058a1fc70c\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.523308 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxh88\" (UniqueName: \"kubernetes.io/projected/10f1a285-434b-4044-9f81-67058a1fc70c-kube-api-access-rxh88\") pod \"10f1a285-434b-4044-9f81-67058a1fc70c\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.523503 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-utilities\") pod \"10f1a285-434b-4044-9f81-67058a1fc70c\" (UID: \"10f1a285-434b-4044-9f81-67058a1fc70c\") " Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.524503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-utilities" (OuterVolumeSpecName: "utilities") pod "10f1a285-434b-4044-9f81-67058a1fc70c" (UID: "10f1a285-434b-4044-9f81-67058a1fc70c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.525070 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.529223 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f1a285-434b-4044-9f81-67058a1fc70c-kube-api-access-rxh88" (OuterVolumeSpecName: "kube-api-access-rxh88") pod "10f1a285-434b-4044-9f81-67058a1fc70c" (UID: "10f1a285-434b-4044-9f81-67058a1fc70c"). InnerVolumeSpecName "kube-api-access-rxh88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.572449 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10f1a285-434b-4044-9f81-67058a1fc70c" (UID: "10f1a285-434b-4044-9f81-67058a1fc70c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.627326 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f1a285-434b-4044-9f81-67058a1fc70c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.627360 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxh88\" (UniqueName: \"kubernetes.io/projected/10f1a285-434b-4044-9f81-67058a1fc70c-kube-api-access-rxh88\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.839684 4726 generic.go:334] "Generic (PLEG): container finished" podID="10f1a285-434b-4044-9f81-67058a1fc70c" containerID="79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87" exitCode=0 Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.839737 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5fs" event={"ID":"10f1a285-434b-4044-9f81-67058a1fc70c","Type":"ContainerDied","Data":"79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87"} Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.839774 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6z5fs" event={"ID":"10f1a285-434b-4044-9f81-67058a1fc70c","Type":"ContainerDied","Data":"0898264973ace674f94b241ac7f6ee422574813af2affb3b1f602d8ead7ed9f8"} Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.839799 4726 scope.go:117] "RemoveContainer" containerID="79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.839999 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6z5fs" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.875703 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6z5fs"] Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.877092 4726 scope.go:117] "RemoveContainer" containerID="e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.885643 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6z5fs"] Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.906096 4726 scope.go:117] "RemoveContainer" containerID="cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.954794 4726 scope.go:117] "RemoveContainer" containerID="79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87" Nov 23 20:32:22 crc kubenswrapper[4726]: E1123 20:32:22.955329 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87\": container with ID starting with 79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87 not found: ID does not exist" containerID="79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.955384 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87"} err="failed to get container status \"79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87\": rpc error: code = NotFound desc = could not find container \"79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87\": container with ID starting with 79939feef0f1f4675fc3ae6c08504b251cc3a52b13a4c302eff6c11eb4724b87 not found: ID does not exist" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.955423 4726 scope.go:117] "RemoveContainer" containerID="e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d" Nov 23 20:32:22 crc kubenswrapper[4726]: E1123 20:32:22.955858 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d\": container with ID starting with e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d not found: ID does not exist" containerID="e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.955929 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d"} err="failed to get container status \"e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d\": rpc error: code = NotFound desc = could not find container \"e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d\": container with ID starting with e72c021b459be56711698a5422e017dc7baec6524ef994b62a9c0f9e548a866d not found: ID does not exist" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.955965 4726 scope.go:117] "RemoveContainer" containerID="cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8" Nov 23 20:32:22 crc kubenswrapper[4726]: E1123 20:32:22.956492 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8\": container with ID starting with cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8 not found: ID does not exist" containerID="cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8" Nov 23 20:32:22 crc kubenswrapper[4726]: I1123 20:32:22.956520 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8"} err="failed to get container status \"cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8\": rpc error: code = NotFound desc = could not find container \"cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8\": container with ID starting with cc1fb1c01040d45851752be02585ec217db6193de0b8d911848197fa833f38a8 not found: ID does not exist" Nov 23 20:32:24 crc kubenswrapper[4726]: I1123 20:32:24.613300 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" path="/var/lib/kubelet/pods/10f1a285-434b-4044-9f81-67058a1fc70c/volumes" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.041567 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-29462"] Nov 23 20:32:28 crc kubenswrapper[4726]: E1123 20:32:28.042414 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" containerName="extract-content" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.042428 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" containerName="extract-content" Nov 23 20:32:28 crc kubenswrapper[4726]: E1123 20:32:28.042445 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" containerName="registry-server" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.042451 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" containerName="registry-server" Nov 23 20:32:28 crc kubenswrapper[4726]: E1123 20:32:28.042461 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="extract-utilities" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.042470 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="extract-utilities" Nov 23 20:32:28 crc kubenswrapper[4726]: E1123 20:32:28.042492 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="extract-content" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.042497 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="extract-content" Nov 23 20:32:28 crc kubenswrapper[4726]: E1123 20:32:28.042505 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="registry-server" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.042511 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="registry-server" Nov 23 20:32:28 crc kubenswrapper[4726]: E1123 20:32:28.042521 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" containerName="extract-utilities" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.042528 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" containerName="extract-utilities" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.042703 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="37272200-e66f-4b21-b337-ff745508d63e" containerName="registry-server" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.042718 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f1a285-434b-4044-9f81-67058a1fc70c" containerName="registry-server" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.043972 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.061390 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29462"] Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.180255 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-catalog-content\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.180841 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjp94\" (UniqueName: \"kubernetes.io/projected/ec76d197-e05b-4282-8cfd-e45006f82bb9-kube-api-access-xjp94\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.181089 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-utilities\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.287695 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-catalog-content\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.287969 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjp94\" (UniqueName: \"kubernetes.io/projected/ec76d197-e05b-4282-8cfd-e45006f82bb9-kube-api-access-xjp94\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.288251 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-utilities\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.289980 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-utilities\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.290314 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-catalog-content\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.323110 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjp94\" (UniqueName: \"kubernetes.io/projected/ec76d197-e05b-4282-8cfd-e45006f82bb9-kube-api-access-xjp94\") pod \"redhat-marketplace-29462\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:28 crc kubenswrapper[4726]: I1123 20:32:28.363193 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:29 crc kubenswrapper[4726]: I1123 20:32:28.821754 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29462"] Nov 23 20:32:29 crc kubenswrapper[4726]: W1123 20:32:28.825305 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec76d197_e05b_4282_8cfd_e45006f82bb9.slice/crio-2f02dcb3f4b1c30c1efd1af5f1d3508cb515f261711bed38dc58a48fc2b082de WatchSource:0}: Error finding container 2f02dcb3f4b1c30c1efd1af5f1d3508cb515f261711bed38dc58a48fc2b082de: Status 404 returned error can't find the container with id 2f02dcb3f4b1c30c1efd1af5f1d3508cb515f261711bed38dc58a48fc2b082de Nov 23 20:32:29 crc kubenswrapper[4726]: I1123 20:32:28.907135 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29462" event={"ID":"ec76d197-e05b-4282-8cfd-e45006f82bb9","Type":"ContainerStarted","Data":"2f02dcb3f4b1c30c1efd1af5f1d3508cb515f261711bed38dc58a48fc2b082de"} Nov 23 20:32:29 crc kubenswrapper[4726]: I1123 20:32:29.916612 4726 generic.go:334] "Generic (PLEG): container finished" podID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerID="053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1" exitCode=0 Nov 23 20:32:29 crc kubenswrapper[4726]: I1123 20:32:29.916665 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29462" event={"ID":"ec76d197-e05b-4282-8cfd-e45006f82bb9","Type":"ContainerDied","Data":"053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1"} Nov 23 20:32:30 crc kubenswrapper[4726]: I1123 20:32:30.927240 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29462" event={"ID":"ec76d197-e05b-4282-8cfd-e45006f82bb9","Type":"ContainerStarted","Data":"c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1"} Nov 23 20:32:31 crc kubenswrapper[4726]: I1123 20:32:31.937176 4726 generic.go:334] "Generic (PLEG): container finished" podID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerID="c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1" exitCode=0 Nov 23 20:32:31 crc kubenswrapper[4726]: I1123 20:32:31.937217 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29462" event={"ID":"ec76d197-e05b-4282-8cfd-e45006f82bb9","Type":"ContainerDied","Data":"c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1"} Nov 23 20:32:32 crc kubenswrapper[4726]: I1123 20:32:32.947189 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29462" event={"ID":"ec76d197-e05b-4282-8cfd-e45006f82bb9","Type":"ContainerStarted","Data":"6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a"} Nov 23 20:32:32 crc kubenswrapper[4726]: I1123 20:32:32.970530 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-29462" podStartSLOduration=2.527037144 podStartE2EDuration="4.970513535s" podCreationTimestamp="2025-11-23 20:32:28 +0000 UTC" firstStartedPulling="2025-11-23 20:32:29.919022357 +0000 UTC m=+1458.068063333" lastFinishedPulling="2025-11-23 20:32:32.362498768 +0000 UTC m=+1460.511539724" observedRunningTime="2025-11-23 20:32:32.968191138 +0000 UTC m=+1461.117232114" watchObservedRunningTime="2025-11-23 20:32:32.970513535 +0000 UTC m=+1461.119554491" Nov 23 20:32:38 crc kubenswrapper[4726]: I1123 20:32:38.363773 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:38 crc kubenswrapper[4726]: I1123 20:32:38.365635 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:38 crc kubenswrapper[4726]: I1123 20:32:38.425070 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:39 crc kubenswrapper[4726]: I1123 20:32:39.043174 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:32:39 crc kubenswrapper[4726]: I1123 20:32:39.043228 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:32:39 crc kubenswrapper[4726]: I1123 20:32:39.065692 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:39 crc kubenswrapper[4726]: I1123 20:32:39.680216 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29462"] Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.023680 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-29462" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerName="registry-server" containerID="cri-o://6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a" gracePeriod=2 Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.531180 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.559045 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-utilities\") pod \"ec76d197-e05b-4282-8cfd-e45006f82bb9\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.559142 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjp94\" (UniqueName: \"kubernetes.io/projected/ec76d197-e05b-4282-8cfd-e45006f82bb9-kube-api-access-xjp94\") pod \"ec76d197-e05b-4282-8cfd-e45006f82bb9\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.559266 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-catalog-content\") pod \"ec76d197-e05b-4282-8cfd-e45006f82bb9\" (UID: \"ec76d197-e05b-4282-8cfd-e45006f82bb9\") " Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.564312 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-utilities" (OuterVolumeSpecName: "utilities") pod "ec76d197-e05b-4282-8cfd-e45006f82bb9" (UID: "ec76d197-e05b-4282-8cfd-e45006f82bb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.569039 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec76d197-e05b-4282-8cfd-e45006f82bb9-kube-api-access-xjp94" (OuterVolumeSpecName: "kube-api-access-xjp94") pod "ec76d197-e05b-4282-8cfd-e45006f82bb9" (UID: "ec76d197-e05b-4282-8cfd-e45006f82bb9"). InnerVolumeSpecName "kube-api-access-xjp94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.597447 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec76d197-e05b-4282-8cfd-e45006f82bb9" (UID: "ec76d197-e05b-4282-8cfd-e45006f82bb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.661788 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.661822 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec76d197-e05b-4282-8cfd-e45006f82bb9-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:41 crc kubenswrapper[4726]: I1123 20:32:41.661835 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjp94\" (UniqueName: \"kubernetes.io/projected/ec76d197-e05b-4282-8cfd-e45006f82bb9-kube-api-access-xjp94\") on node \"crc\" DevicePath \"\"" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.036977 4726 generic.go:334] "Generic (PLEG): container finished" podID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerID="6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a" exitCode=0 Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.037054 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29462" event={"ID":"ec76d197-e05b-4282-8cfd-e45006f82bb9","Type":"ContainerDied","Data":"6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a"} Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.037129 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29462" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.037172 4726 scope.go:117] "RemoveContainer" containerID="6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.037148 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29462" event={"ID":"ec76d197-e05b-4282-8cfd-e45006f82bb9","Type":"ContainerDied","Data":"2f02dcb3f4b1c30c1efd1af5f1d3508cb515f261711bed38dc58a48fc2b082de"} Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.093458 4726 scope.go:117] "RemoveContainer" containerID="c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.093985 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29462"] Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.110135 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-29462"] Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.119352 4726 scope.go:117] "RemoveContainer" containerID="053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.165059 4726 scope.go:117] "RemoveContainer" containerID="6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a" Nov 23 20:32:42 crc kubenswrapper[4726]: E1123 20:32:42.165603 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a\": container with ID starting with 6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a not found: ID does not exist" containerID="6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.165644 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a"} err="failed to get container status \"6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a\": rpc error: code = NotFound desc = could not find container \"6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a\": container with ID starting with 6105af7b6036492cb6486e007beadf6bb5ec2ab59e9cc918995be2081cd5990a not found: ID does not exist" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.165670 4726 scope.go:117] "RemoveContainer" containerID="c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1" Nov 23 20:32:42 crc kubenswrapper[4726]: E1123 20:32:42.166102 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1\": container with ID starting with c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1 not found: ID does not exist" containerID="c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.166137 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1"} err="failed to get container status \"c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1\": rpc error: code = NotFound desc = could not find container \"c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1\": container with ID starting with c18a47ba2db1af7ba58b03c832b771aa10d6f1bf4ab6d34b2d10d628b8c2dbf1 not found: ID does not exist" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.166159 4726 scope.go:117] "RemoveContainer" containerID="053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1" Nov 23 20:32:42 crc kubenswrapper[4726]: E1123 20:32:42.166635 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1\": container with ID starting with 053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1 not found: ID does not exist" containerID="053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.166706 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1"} err="failed to get container status \"053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1\": rpc error: code = NotFound desc = could not find container \"053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1\": container with ID starting with 053bf59d56a8c0a2949e470ff3967d2be4c678b1945d9c0f4205002aaa0203e1 not found: ID does not exist" Nov 23 20:32:42 crc kubenswrapper[4726]: I1123 20:32:42.601575 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" path="/var/lib/kubelet/pods/ec76d197-e05b-4282-8cfd-e45006f82bb9/volumes" Nov 23 20:33:09 crc kubenswrapper[4726]: I1123 20:33:09.043667 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:33:09 crc kubenswrapper[4726]: I1123 20:33:09.045139 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:33:17 crc kubenswrapper[4726]: I1123 20:33:17.198253 4726 scope.go:117] "RemoveContainer" containerID="10ab5f69ff760f73f6a759bee2a03a28da55c14b547d6fa0d20f9365c38ceb00" Nov 23 20:33:17 crc kubenswrapper[4726]: I1123 20:33:17.220730 4726 scope.go:117] "RemoveContainer" containerID="9df51064fc9d4ce90560cb54e4dcfaf2d9ec35c60f07f233fae2fc11cd9b013c" Nov 23 20:33:17 crc kubenswrapper[4726]: I1123 20:33:17.244847 4726 scope.go:117] "RemoveContainer" containerID="1d19b5df6e99c0abcd89710424fc852a724937af1d0b2047b2044e4c7a25f1ce" Nov 23 20:33:17 crc kubenswrapper[4726]: I1123 20:33:17.277253 4726 scope.go:117] "RemoveContainer" containerID="6fdcd167e72753d7df9099060cd79c81fe26cdd50cbc71f1fea8960e030d3c08" Nov 23 20:33:17 crc kubenswrapper[4726]: I1123 20:33:17.298410 4726 scope.go:117] "RemoveContainer" containerID="c1c44993720e8bfe84abc72ead753ff54c098bc2f4c92bef50223afbc08af850" Nov 23 20:33:30 crc kubenswrapper[4726]: E1123 20:33:30.277391 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ff7e920_cb8b_453e_82cb_454e59325a43.slice/crio-conmon-740fcead03fe7884ee89023a0f11ad53bdc4e850757cd8a8a92051626d94f729.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ff7e920_cb8b_453e_82cb_454e59325a43.slice/crio-740fcead03fe7884ee89023a0f11ad53bdc4e850757cd8a8a92051626d94f729.scope\": RecentStats: unable to find data in memory cache]" Nov 23 20:33:30 crc kubenswrapper[4726]: I1123 20:33:30.539531 4726 generic.go:334] "Generic (PLEG): container finished" podID="0ff7e920-cb8b-453e-82cb-454e59325a43" containerID="740fcead03fe7884ee89023a0f11ad53bdc4e850757cd8a8a92051626d94f729" exitCode=0 Nov 23 20:33:30 crc kubenswrapper[4726]: I1123 20:33:30.540675 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" event={"ID":"0ff7e920-cb8b-453e-82cb-454e59325a43","Type":"ContainerDied","Data":"740fcead03fe7884ee89023a0f11ad53bdc4e850757cd8a8a92051626d94f729"} Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.048335 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.176192 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-bootstrap-combined-ca-bundle\") pod \"0ff7e920-cb8b-453e-82cb-454e59325a43\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.176543 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-inventory\") pod \"0ff7e920-cb8b-453e-82cb-454e59325a43\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.176603 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-ssh-key\") pod \"0ff7e920-cb8b-453e-82cb-454e59325a43\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.176658 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcxzp\" (UniqueName: \"kubernetes.io/projected/0ff7e920-cb8b-453e-82cb-454e59325a43-kube-api-access-bcxzp\") pod \"0ff7e920-cb8b-453e-82cb-454e59325a43\" (UID: \"0ff7e920-cb8b-453e-82cb-454e59325a43\") " Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.185516 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ff7e920-cb8b-453e-82cb-454e59325a43-kube-api-access-bcxzp" (OuterVolumeSpecName: "kube-api-access-bcxzp") pod "0ff7e920-cb8b-453e-82cb-454e59325a43" (UID: "0ff7e920-cb8b-453e-82cb-454e59325a43"). InnerVolumeSpecName "kube-api-access-bcxzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.187134 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0ff7e920-cb8b-453e-82cb-454e59325a43" (UID: "0ff7e920-cb8b-453e-82cb-454e59325a43"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.207825 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0ff7e920-cb8b-453e-82cb-454e59325a43" (UID: "0ff7e920-cb8b-453e-82cb-454e59325a43"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.233843 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-inventory" (OuterVolumeSpecName: "inventory") pod "0ff7e920-cb8b-453e-82cb-454e59325a43" (UID: "0ff7e920-cb8b-453e-82cb-454e59325a43"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.280358 4726 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.280399 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.280411 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ff7e920-cb8b-453e-82cb-454e59325a43-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.280424 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcxzp\" (UniqueName: \"kubernetes.io/projected/0ff7e920-cb8b-453e-82cb-454e59325a43-kube-api-access-bcxzp\") on node \"crc\" DevicePath \"\"" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.566974 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" event={"ID":"0ff7e920-cb8b-453e-82cb-454e59325a43","Type":"ContainerDied","Data":"925fcb30dec6c94fc33872d2d21a7522fa7fe1dacb1dbb373b44c99d1338b0ce"} Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.567284 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="925fcb30dec6c94fc33872d2d21a7522fa7fe1dacb1dbb373b44c99d1338b0ce" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.567075 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.686543 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59"] Nov 23 20:33:32 crc kubenswrapper[4726]: E1123 20:33:32.686918 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerName="extract-content" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.686935 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerName="extract-content" Nov 23 20:33:32 crc kubenswrapper[4726]: E1123 20:33:32.686958 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerName="extract-utilities" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.686965 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerName="extract-utilities" Nov 23 20:33:32 crc kubenswrapper[4726]: E1123 20:33:32.686980 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff7e920-cb8b-453e-82cb-454e59325a43" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.686988 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff7e920-cb8b-453e-82cb-454e59325a43" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 20:33:32 crc kubenswrapper[4726]: E1123 20:33:32.687000 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerName="registry-server" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.687006 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerName="registry-server" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.687545 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec76d197-e05b-4282-8cfd-e45006f82bb9" containerName="registry-server" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.687578 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ff7e920-cb8b-453e-82cb-454e59325a43" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.688380 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.691660 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.691823 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.692203 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.692587 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.727928 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59"] Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.826692 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.827150 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.827218 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44w4l\" (UniqueName: \"kubernetes.io/projected/acbca385-261a-4dd2-98e3-602ee283e043-kube-api-access-44w4l\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.928960 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.929312 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.929685 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44w4l\" (UniqueName: \"kubernetes.io/projected/acbca385-261a-4dd2-98e3-602ee283e043-kube-api-access-44w4l\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.936402 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.938087 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:32 crc kubenswrapper[4726]: I1123 20:33:32.953976 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44w4l\" (UniqueName: \"kubernetes.io/projected/acbca385-261a-4dd2-98e3-602ee283e043-kube-api-access-44w4l\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jnt59\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:33 crc kubenswrapper[4726]: I1123 20:33:33.012917 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:33:33 crc kubenswrapper[4726]: I1123 20:33:33.383348 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59"] Nov 23 20:33:33 crc kubenswrapper[4726]: I1123 20:33:33.579825 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" event={"ID":"acbca385-261a-4dd2-98e3-602ee283e043","Type":"ContainerStarted","Data":"f8be9b39ad99eddb4cea29bded5b9a65144f4018a04a9993735feefed8501ade"} Nov 23 20:33:34 crc kubenswrapper[4726]: I1123 20:33:34.605447 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" event={"ID":"acbca385-261a-4dd2-98e3-602ee283e043","Type":"ContainerStarted","Data":"8cd5808554f84ac6456cde393aed0c648e4a4e329828651bc572e8dcc884b9a6"} Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.043431 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.045069 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.045223 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.046407 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.046597 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" gracePeriod=600 Nov 23 20:33:39 crc kubenswrapper[4726]: E1123 20:33:39.179230 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.657499 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" exitCode=0 Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.657548 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa"} Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.657582 4726 scope.go:117] "RemoveContainer" containerID="828799a4d998638471568a30864015defe9240aef8698bc3b256f460fe153d23" Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.658263 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:33:39 crc kubenswrapper[4726]: E1123 20:33:39.658611 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:33:39 crc kubenswrapper[4726]: I1123 20:33:39.682935 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" podStartSLOduration=7.1671236369999995 podStartE2EDuration="7.682912167s" podCreationTimestamp="2025-11-23 20:33:32 +0000 UTC" firstStartedPulling="2025-11-23 20:33:33.394508991 +0000 UTC m=+1521.543549947" lastFinishedPulling="2025-11-23 20:33:33.910297521 +0000 UTC m=+1522.059338477" observedRunningTime="2025-11-23 20:33:34.618374709 +0000 UTC m=+1522.767415665" watchObservedRunningTime="2025-11-23 20:33:39.682912167 +0000 UTC m=+1527.831953123" Nov 23 20:33:53 crc kubenswrapper[4726]: I1123 20:33:53.589256 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:33:53 crc kubenswrapper[4726]: E1123 20:33:53.590486 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:34:01 crc kubenswrapper[4726]: I1123 20:34:01.055559 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4d43-account-create-jvb7s"] Nov 23 20:34:01 crc kubenswrapper[4726]: I1123 20:34:01.066044 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-4k65v"] Nov 23 20:34:01 crc kubenswrapper[4726]: I1123 20:34:01.076363 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-x7cvm"] Nov 23 20:34:01 crc kubenswrapper[4726]: I1123 20:34:01.084133 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4d43-account-create-jvb7s"] Nov 23 20:34:01 crc kubenswrapper[4726]: I1123 20:34:01.091511 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-4k65v"] Nov 23 20:34:01 crc kubenswrapper[4726]: I1123 20:34:01.097901 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-x7cvm"] Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.051077 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-118c-account-create-jmbld"] Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.065973 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-eacb-account-create-wn7kh"] Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.075084 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-118c-account-create-jmbld"] Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.082937 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-eacb-account-create-wn7kh"] Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.089028 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-d6nsj"] Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.095489 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-d6nsj"] Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.601532 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fd3ee85-5046-4edf-9d13-413525f0752e" path="/var/lib/kubelet/pods/3fd3ee85-5046-4edf-9d13-413525f0752e/volumes" Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.602139 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6643a95f-f556-4e3e-9a45-faab4f4772b6" path="/var/lib/kubelet/pods/6643a95f-f556-4e3e-9a45-faab4f4772b6/volumes" Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.602683 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c50815e-46f6-4f70-be21-4c2c9f98bf35" path="/var/lib/kubelet/pods/7c50815e-46f6-4f70-be21-4c2c9f98bf35/volumes" Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.603238 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0dacd12-8219-4c35-8d8b-3903cfb98c63" path="/var/lib/kubelet/pods/e0dacd12-8219-4c35-8d8b-3903cfb98c63/volumes" Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.604243 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eba1f02c-d02a-4c20-a09f-b3292cd23a0d" path="/var/lib/kubelet/pods/eba1f02c-d02a-4c20-a09f-b3292cd23a0d/volumes" Nov 23 20:34:02 crc kubenswrapper[4726]: I1123 20:34:02.604790 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eee8014b-ecc7-4756-967e-3da08ee9192e" path="/var/lib/kubelet/pods/eee8014b-ecc7-4756-967e-3da08ee9192e/volumes" Nov 23 20:34:08 crc kubenswrapper[4726]: I1123 20:34:08.589670 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:34:08 crc kubenswrapper[4726]: E1123 20:34:08.590690 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:34:17 crc kubenswrapper[4726]: I1123 20:34:17.400568 4726 scope.go:117] "RemoveContainer" containerID="c6514513a7733b2cf84485c1c6ceea1579f973052e8a50517556f92a36fcffda" Nov 23 20:34:17 crc kubenswrapper[4726]: I1123 20:34:17.433169 4726 scope.go:117] "RemoveContainer" containerID="b88d48fea82f8ade138ebc58abe6c0feac1624b7f966409096d9e0d3985823a1" Nov 23 20:34:17 crc kubenswrapper[4726]: I1123 20:34:17.477103 4726 scope.go:117] "RemoveContainer" containerID="c6853c4b4d49170bfa1661ced9b933946d1faccf6560802174cb7ed1ebf3e480" Nov 23 20:34:17 crc kubenswrapper[4726]: I1123 20:34:17.522410 4726 scope.go:117] "RemoveContainer" containerID="4cd02756ae106783e889092006a66e574deeb783f7245b06539f4e0e2909f8f0" Nov 23 20:34:17 crc kubenswrapper[4726]: I1123 20:34:17.560823 4726 scope.go:117] "RemoveContainer" containerID="5b12cec731b054532248b400473ac69038e166fd7a10d4df07e78851dc36ecda" Nov 23 20:34:17 crc kubenswrapper[4726]: I1123 20:34:17.598495 4726 scope.go:117] "RemoveContainer" containerID="861df19309cc381e13b50ead4bc23c6863d43f94befe38d31db2119f945c3c27" Nov 23 20:34:20 crc kubenswrapper[4726]: I1123 20:34:20.589973 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:34:20 crc kubenswrapper[4726]: E1123 20:34:20.591714 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:34:26 crc kubenswrapper[4726]: I1123 20:34:26.040746 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-fffk2"] Nov 23 20:34:26 crc kubenswrapper[4726]: I1123 20:34:26.050239 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-fffk2"] Nov 23 20:34:26 crc kubenswrapper[4726]: I1123 20:34:26.623182 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6877e1a3-06c3-4619-9ec1-61cecf78c66f" path="/var/lib/kubelet/pods/6877e1a3-06c3-4619-9ec1-61cecf78c66f/volumes" Nov 23 20:34:34 crc kubenswrapper[4726]: I1123 20:34:34.589835 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:34:34 crc kubenswrapper[4726]: E1123 20:34:34.590810 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:34:40 crc kubenswrapper[4726]: I1123 20:34:40.058134 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5d0d-account-create-tn2pk"] Nov 23 20:34:40 crc kubenswrapper[4726]: I1123 20:34:40.067627 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-m4fr4"] Nov 23 20:34:40 crc kubenswrapper[4726]: I1123 20:34:40.076084 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-m4fr4"] Nov 23 20:34:40 crc kubenswrapper[4726]: I1123 20:34:40.085244 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5d0d-account-create-tn2pk"] Nov 23 20:34:40 crc kubenswrapper[4726]: I1123 20:34:40.598384 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd" path="/var/lib/kubelet/pods/146e95c1-b7e2-4ebc-880d-bd89cf7cb0cd/volumes" Nov 23 20:34:40 crc kubenswrapper[4726]: I1123 20:34:40.599046 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b7c9ec1-e0d0-48cb-bfec-510f28a2a234" path="/var/lib/kubelet/pods/7b7c9ec1-e0d0-48cb-bfec-510f28a2a234/volumes" Nov 23 20:34:43 crc kubenswrapper[4726]: I1123 20:34:43.040035 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-psjjc"] Nov 23 20:34:43 crc kubenswrapper[4726]: I1123 20:34:43.061192 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-769a-account-create-qj28b"] Nov 23 20:34:43 crc kubenswrapper[4726]: I1123 20:34:43.071644 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-769a-account-create-qj28b"] Nov 23 20:34:43 crc kubenswrapper[4726]: I1123 20:34:43.078767 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-psjjc"] Nov 23 20:34:43 crc kubenswrapper[4726]: I1123 20:34:43.086467 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b6a0-account-create-2zr8p"] Nov 23 20:34:43 crc kubenswrapper[4726]: I1123 20:34:43.093525 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-x7d9j"] Nov 23 20:34:43 crc kubenswrapper[4726]: I1123 20:34:43.099985 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-x7d9j"] Nov 23 20:34:43 crc kubenswrapper[4726]: I1123 20:34:43.106104 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b6a0-account-create-2zr8p"] Nov 23 20:34:44 crc kubenswrapper[4726]: I1123 20:34:44.602431 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c555255-c599-4b91-9a76-c860a2c99228" path="/var/lib/kubelet/pods/1c555255-c599-4b91-9a76-c860a2c99228/volumes" Nov 23 20:34:44 crc kubenswrapper[4726]: I1123 20:34:44.603262 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51cac0d0-29b4-4867-9279-2142c34e613c" path="/var/lib/kubelet/pods/51cac0d0-29b4-4867-9279-2142c34e613c/volumes" Nov 23 20:34:44 crc kubenswrapper[4726]: I1123 20:34:44.604094 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c555bb2-ca8a-412e-852f-8f928feeab8d" path="/var/lib/kubelet/pods/5c555bb2-ca8a-412e-852f-8f928feeab8d/volumes" Nov 23 20:34:44 crc kubenswrapper[4726]: I1123 20:34:44.604804 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa225da7-111f-449f-8e6e-d2a73bb39a8f" path="/var/lib/kubelet/pods/aa225da7-111f-449f-8e6e-d2a73bb39a8f/volumes" Nov 23 20:34:49 crc kubenswrapper[4726]: I1123 20:34:49.071334 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-rnkvw"] Nov 23 20:34:49 crc kubenswrapper[4726]: I1123 20:34:49.081317 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-rnkvw"] Nov 23 20:34:49 crc kubenswrapper[4726]: I1123 20:34:49.588672 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:34:49 crc kubenswrapper[4726]: E1123 20:34:49.589132 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:34:50 crc kubenswrapper[4726]: I1123 20:34:50.599412 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d16b97d-7a0e-4312-94c8-f451d9e66d2b" path="/var/lib/kubelet/pods/2d16b97d-7a0e-4312-94c8-f451d9e66d2b/volumes" Nov 23 20:34:56 crc kubenswrapper[4726]: I1123 20:34:56.520422 4726 generic.go:334] "Generic (PLEG): container finished" podID="acbca385-261a-4dd2-98e3-602ee283e043" containerID="8cd5808554f84ac6456cde393aed0c648e4a4e329828651bc572e8dcc884b9a6" exitCode=0 Nov 23 20:34:56 crc kubenswrapper[4726]: I1123 20:34:56.520512 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" event={"ID":"acbca385-261a-4dd2-98e3-602ee283e043","Type":"ContainerDied","Data":"8cd5808554f84ac6456cde393aed0c648e4a4e329828651bc572e8dcc884b9a6"} Nov 23 20:34:57 crc kubenswrapper[4726]: I1123 20:34:57.959688 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.041317 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44w4l\" (UniqueName: \"kubernetes.io/projected/acbca385-261a-4dd2-98e3-602ee283e043-kube-api-access-44w4l\") pod \"acbca385-261a-4dd2-98e3-602ee283e043\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.041392 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-ssh-key\") pod \"acbca385-261a-4dd2-98e3-602ee283e043\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.041504 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-inventory\") pod \"acbca385-261a-4dd2-98e3-602ee283e043\" (UID: \"acbca385-261a-4dd2-98e3-602ee283e043\") " Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.056137 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acbca385-261a-4dd2-98e3-602ee283e043-kube-api-access-44w4l" (OuterVolumeSpecName: "kube-api-access-44w4l") pod "acbca385-261a-4dd2-98e3-602ee283e043" (UID: "acbca385-261a-4dd2-98e3-602ee283e043"). InnerVolumeSpecName "kube-api-access-44w4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.071181 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "acbca385-261a-4dd2-98e3-602ee283e043" (UID: "acbca385-261a-4dd2-98e3-602ee283e043"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.077879 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-inventory" (OuterVolumeSpecName: "inventory") pod "acbca385-261a-4dd2-98e3-602ee283e043" (UID: "acbca385-261a-4dd2-98e3-602ee283e043"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.143723 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.143756 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44w4l\" (UniqueName: \"kubernetes.io/projected/acbca385-261a-4dd2-98e3-602ee283e043-kube-api-access-44w4l\") on node \"crc\" DevicePath \"\"" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.143766 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acbca385-261a-4dd2-98e3-602ee283e043-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.542849 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" event={"ID":"acbca385-261a-4dd2-98e3-602ee283e043","Type":"ContainerDied","Data":"f8be9b39ad99eddb4cea29bded5b9a65144f4018a04a9993735feefed8501ade"} Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.542949 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8be9b39ad99eddb4cea29bded5b9a65144f4018a04a9993735feefed8501ade" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.543352 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.655108 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d"] Nov 23 20:34:58 crc kubenswrapper[4726]: E1123 20:34:58.655610 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbca385-261a-4dd2-98e3-602ee283e043" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.655631 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbca385-261a-4dd2-98e3-602ee283e043" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.655910 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="acbca385-261a-4dd2-98e3-602ee283e043" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.656724 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.663003 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.663398 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.663467 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.664049 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.666191 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d"] Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.754048 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.754139 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvz7s\" (UniqueName: \"kubernetes.io/projected/4e36024a-9586-439f-b551-d5baba64775b-kube-api-access-vvz7s\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.754189 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.855920 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.856191 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvz7s\" (UniqueName: \"kubernetes.io/projected/4e36024a-9586-439f-b551-d5baba64775b-kube-api-access-vvz7s\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.856299 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.859654 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.867848 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:58 crc kubenswrapper[4726]: I1123 20:34:58.885890 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvz7s\" (UniqueName: \"kubernetes.io/projected/4e36024a-9586-439f-b551-d5baba64775b-kube-api-access-vvz7s\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-44l9d\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:59 crc kubenswrapper[4726]: I1123 20:34:58.993411 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:34:59 crc kubenswrapper[4726]: I1123 20:34:59.529478 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d"] Nov 23 20:34:59 crc kubenswrapper[4726]: W1123 20:34:59.537577 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e36024a_9586_439f_b551_d5baba64775b.slice/crio-a4edfd463c77744e4db593ca6c84bcc096659981bd2e70c0512ab681a685adac WatchSource:0}: Error finding container a4edfd463c77744e4db593ca6c84bcc096659981bd2e70c0512ab681a685adac: Status 404 returned error can't find the container with id a4edfd463c77744e4db593ca6c84bcc096659981bd2e70c0512ab681a685adac Nov 23 20:34:59 crc kubenswrapper[4726]: I1123 20:34:59.542544 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 20:34:59 crc kubenswrapper[4726]: I1123 20:34:59.557129 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" event={"ID":"4e36024a-9586-439f-b551-d5baba64775b","Type":"ContainerStarted","Data":"a4edfd463c77744e4db593ca6c84bcc096659981bd2e70c0512ab681a685adac"} Nov 23 20:35:00 crc kubenswrapper[4726]: I1123 20:35:00.572376 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" event={"ID":"4e36024a-9586-439f-b551-d5baba64775b","Type":"ContainerStarted","Data":"2553283fdb55c90da0ab0b30e350932a587049af4077cf26d70079a6397a7ad2"} Nov 23 20:35:00 crc kubenswrapper[4726]: I1123 20:35:00.609460 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" podStartSLOduration=2.105686592 podStartE2EDuration="2.609430403s" podCreationTimestamp="2025-11-23 20:34:58 +0000 UTC" firstStartedPulling="2025-11-23 20:34:59.542258689 +0000 UTC m=+1607.691299655" lastFinishedPulling="2025-11-23 20:35:00.0460025 +0000 UTC m=+1608.195043466" observedRunningTime="2025-11-23 20:35:00.601726425 +0000 UTC m=+1608.750767451" watchObservedRunningTime="2025-11-23 20:35:00.609430403 +0000 UTC m=+1608.758471399" Nov 23 20:35:02 crc kubenswrapper[4726]: I1123 20:35:02.619089 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:35:02 crc kubenswrapper[4726]: E1123 20:35:02.620136 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:35:06 crc kubenswrapper[4726]: I1123 20:35:06.636021 4726 generic.go:334] "Generic (PLEG): container finished" podID="4e36024a-9586-439f-b551-d5baba64775b" containerID="2553283fdb55c90da0ab0b30e350932a587049af4077cf26d70079a6397a7ad2" exitCode=0 Nov 23 20:35:06 crc kubenswrapper[4726]: I1123 20:35:06.636135 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" event={"ID":"4e36024a-9586-439f-b551-d5baba64775b","Type":"ContainerDied","Data":"2553283fdb55c90da0ab0b30e350932a587049af4077cf26d70079a6397a7ad2"} Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.095411 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.235593 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvz7s\" (UniqueName: \"kubernetes.io/projected/4e36024a-9586-439f-b551-d5baba64775b-kube-api-access-vvz7s\") pod \"4e36024a-9586-439f-b551-d5baba64775b\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.235808 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-ssh-key\") pod \"4e36024a-9586-439f-b551-d5baba64775b\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.235905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-inventory\") pod \"4e36024a-9586-439f-b551-d5baba64775b\" (UID: \"4e36024a-9586-439f-b551-d5baba64775b\") " Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.242742 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e36024a-9586-439f-b551-d5baba64775b-kube-api-access-vvz7s" (OuterVolumeSpecName: "kube-api-access-vvz7s") pod "4e36024a-9586-439f-b551-d5baba64775b" (UID: "4e36024a-9586-439f-b551-d5baba64775b"). InnerVolumeSpecName "kube-api-access-vvz7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.265016 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4e36024a-9586-439f-b551-d5baba64775b" (UID: "4e36024a-9586-439f-b551-d5baba64775b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.278162 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-inventory" (OuterVolumeSpecName: "inventory") pod "4e36024a-9586-439f-b551-d5baba64775b" (UID: "4e36024a-9586-439f-b551-d5baba64775b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.337983 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.338821 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvz7s\" (UniqueName: \"kubernetes.io/projected/4e36024a-9586-439f-b551-d5baba64775b-kube-api-access-vvz7s\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.338926 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e36024a-9586-439f-b551-d5baba64775b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.669823 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" event={"ID":"4e36024a-9586-439f-b551-d5baba64775b","Type":"ContainerDied","Data":"a4edfd463c77744e4db593ca6c84bcc096659981bd2e70c0512ab681a685adac"} Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.669905 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.669911 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4edfd463c77744e4db593ca6c84bcc096659981bd2e70c0512ab681a685adac" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.742737 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k"] Nov 23 20:35:08 crc kubenswrapper[4726]: E1123 20:35:08.744084 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e36024a-9586-439f-b551-d5baba64775b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.744116 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e36024a-9586-439f-b551-d5baba64775b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.744355 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e36024a-9586-439f-b551-d5baba64775b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.745236 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.748535 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.748707 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.748855 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.749748 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.750654 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k"] Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.848236 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.849018 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftpv8\" (UniqueName: \"kubernetes.io/projected/70977db3-0e6c-4718-bebe-1d14ad030b9d-kube-api-access-ftpv8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.849238 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.950632 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftpv8\" (UniqueName: \"kubernetes.io/projected/70977db3-0e6c-4718-bebe-1d14ad030b9d-kube-api-access-ftpv8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.950751 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.950844 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.956786 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.967965 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:08 crc kubenswrapper[4726]: I1123 20:35:08.972961 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftpv8\" (UniqueName: \"kubernetes.io/projected/70977db3-0e6c-4718-bebe-1d14ad030b9d-kube-api-access-ftpv8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pm78k\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:09 crc kubenswrapper[4726]: I1123 20:35:09.064120 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:09 crc kubenswrapper[4726]: I1123 20:35:09.614410 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k"] Nov 23 20:35:09 crc kubenswrapper[4726]: I1123 20:35:09.679007 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" event={"ID":"70977db3-0e6c-4718-bebe-1d14ad030b9d","Type":"ContainerStarted","Data":"6918c48b51dcb5807ff342bf93892bb757f7b3f73a56542cf69caec1ce7ab801"} Nov 23 20:35:10 crc kubenswrapper[4726]: I1123 20:35:10.690479 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" event={"ID":"70977db3-0e6c-4718-bebe-1d14ad030b9d","Type":"ContainerStarted","Data":"6279b72420a8ef5e3fd08686627f37edef6851aad393e3a83b76912dba0ce494"} Nov 23 20:35:10 crc kubenswrapper[4726]: I1123 20:35:10.709924 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" podStartSLOduration=2.303061136 podStartE2EDuration="2.709902041s" podCreationTimestamp="2025-11-23 20:35:08 +0000 UTC" firstStartedPulling="2025-11-23 20:35:09.634933547 +0000 UTC m=+1617.783974503" lastFinishedPulling="2025-11-23 20:35:10.041774452 +0000 UTC m=+1618.190815408" observedRunningTime="2025-11-23 20:35:10.708388278 +0000 UTC m=+1618.857429244" watchObservedRunningTime="2025-11-23 20:35:10.709902041 +0000 UTC m=+1618.858942997" Nov 23 20:35:15 crc kubenswrapper[4726]: I1123 20:35:15.589708 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:35:15 crc kubenswrapper[4726]: E1123 20:35:15.590672 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:35:17 crc kubenswrapper[4726]: I1123 20:35:17.760407 4726 scope.go:117] "RemoveContainer" containerID="13d56dbd9d7aa9fc206ba2b30c26eadececfea18f4c59621c62b00bf1dcd213f" Nov 23 20:35:17 crc kubenswrapper[4726]: I1123 20:35:17.788441 4726 scope.go:117] "RemoveContainer" containerID="7a6b3d215f9277e5a0e3dd2b210271a307ecdc0c5926dfc2f7c4f324af667aad" Nov 23 20:35:17 crc kubenswrapper[4726]: I1123 20:35:17.861835 4726 scope.go:117] "RemoveContainer" containerID="c57a772b2139f336b20a1afbc000cb7a34e5bbd0db983bc5f3c46f84f843e7bc" Nov 23 20:35:17 crc kubenswrapper[4726]: I1123 20:35:17.930532 4726 scope.go:117] "RemoveContainer" containerID="bd3dfe55a7358e627b47808f2d18e797ff6107b0a0ba799755b44a27b2c8600f" Nov 23 20:35:17 crc kubenswrapper[4726]: I1123 20:35:17.951714 4726 scope.go:117] "RemoveContainer" containerID="b0f11e8a5604fa2461a74c448c4711802b50b13f152ad1cec2f187d7ced542d6" Nov 23 20:35:17 crc kubenswrapper[4726]: I1123 20:35:17.985195 4726 scope.go:117] "RemoveContainer" containerID="019216e2312cb4f4fb42b0f405f65053f655bca0deaddf59b331b79a020364fb" Nov 23 20:35:18 crc kubenswrapper[4726]: I1123 20:35:18.024257 4726 scope.go:117] "RemoveContainer" containerID="b4ff4cd21e8d372185cb63ca270970b780134976af3311ba7521402de7aeba88" Nov 23 20:35:18 crc kubenswrapper[4726]: I1123 20:35:18.057338 4726 scope.go:117] "RemoveContainer" containerID="a44955defed0c124ad62cd79470d7c1fb78def6a01bfebecf3c07c0bde9d2bdd" Nov 23 20:35:23 crc kubenswrapper[4726]: I1123 20:35:23.073544 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-gcxk9"] Nov 23 20:35:23 crc kubenswrapper[4726]: I1123 20:35:23.089105 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4tm6n"] Nov 23 20:35:23 crc kubenswrapper[4726]: I1123 20:35:23.096899 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-gcxk9"] Nov 23 20:35:23 crc kubenswrapper[4726]: I1123 20:35:23.108444 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4tm6n"] Nov 23 20:35:23 crc kubenswrapper[4726]: I1123 20:35:23.116257 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qk592"] Nov 23 20:35:23 crc kubenswrapper[4726]: I1123 20:35:23.123537 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qk592"] Nov 23 20:35:24 crc kubenswrapper[4726]: I1123 20:35:24.603772 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c3f7da5-afd6-44b1-b402-f3de8c39b518" path="/var/lib/kubelet/pods/0c3f7da5-afd6-44b1-b402-f3de8c39b518/volumes" Nov 23 20:35:24 crc kubenswrapper[4726]: I1123 20:35:24.605261 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="595377fa-b0c8-440a-a6f9-4bce1ce40352" path="/var/lib/kubelet/pods/595377fa-b0c8-440a-a6f9-4bce1ce40352/volumes" Nov 23 20:35:24 crc kubenswrapper[4726]: I1123 20:35:24.606032 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4" path="/var/lib/kubelet/pods/c9c66f3f-a5c5-4e41-b1a4-6da9d0e18cf4/volumes" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.017324 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ntj8n"] Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.019513 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.112600 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ntj8n"] Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.149608 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cnq9\" (UniqueName: \"kubernetes.io/projected/30b465cd-2700-4659-9627-42833b738211-kube-api-access-2cnq9\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.150065 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-catalog-content\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.150221 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-utilities\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.252572 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-utilities\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.252714 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cnq9\" (UniqueName: \"kubernetes.io/projected/30b465cd-2700-4659-9627-42833b738211-kube-api-access-2cnq9\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.252763 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-catalog-content\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.253566 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-catalog-content\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.255012 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-utilities\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.276113 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cnq9\" (UniqueName: \"kubernetes.io/projected/30b465cd-2700-4659-9627-42833b738211-kube-api-access-2cnq9\") pod \"certified-operators-ntj8n\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.343469 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.589261 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:35:27 crc kubenswrapper[4726]: E1123 20:35:27.590046 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.814225 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ntj8n"] Nov 23 20:35:27 crc kubenswrapper[4726]: I1123 20:35:27.879092 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ntj8n" event={"ID":"30b465cd-2700-4659-9627-42833b738211","Type":"ContainerStarted","Data":"f134aac5aadb35871132015405e41fc80a297d2a2353f9e04f305150b82a1d99"} Nov 23 20:35:28 crc kubenswrapper[4726]: I1123 20:35:28.888271 4726 generic.go:334] "Generic (PLEG): container finished" podID="30b465cd-2700-4659-9627-42833b738211" containerID="f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9" exitCode=0 Nov 23 20:35:28 crc kubenswrapper[4726]: I1123 20:35:28.888514 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ntj8n" event={"ID":"30b465cd-2700-4659-9627-42833b738211","Type":"ContainerDied","Data":"f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9"} Nov 23 20:35:29 crc kubenswrapper[4726]: I1123 20:35:29.897964 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ntj8n" event={"ID":"30b465cd-2700-4659-9627-42833b738211","Type":"ContainerStarted","Data":"87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a"} Nov 23 20:35:30 crc kubenswrapper[4726]: I1123 20:35:30.907280 4726 generic.go:334] "Generic (PLEG): container finished" podID="30b465cd-2700-4659-9627-42833b738211" containerID="87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a" exitCode=0 Nov 23 20:35:30 crc kubenswrapper[4726]: I1123 20:35:30.907324 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ntj8n" event={"ID":"30b465cd-2700-4659-9627-42833b738211","Type":"ContainerDied","Data":"87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a"} Nov 23 20:35:31 crc kubenswrapper[4726]: I1123 20:35:31.918185 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ntj8n" event={"ID":"30b465cd-2700-4659-9627-42833b738211","Type":"ContainerStarted","Data":"4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d"} Nov 23 20:35:31 crc kubenswrapper[4726]: I1123 20:35:31.944984 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ntj8n" podStartSLOduration=3.415938841 podStartE2EDuration="5.944963409s" podCreationTimestamp="2025-11-23 20:35:26 +0000 UTC" firstStartedPulling="2025-11-23 20:35:28.889816915 +0000 UTC m=+1637.038857871" lastFinishedPulling="2025-11-23 20:35:31.418841483 +0000 UTC m=+1639.567882439" observedRunningTime="2025-11-23 20:35:31.936249942 +0000 UTC m=+1640.085290928" watchObservedRunningTime="2025-11-23 20:35:31.944963409 +0000 UTC m=+1640.094004375" Nov 23 20:35:33 crc kubenswrapper[4726]: I1123 20:35:33.040224 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-s7rvr"] Nov 23 20:35:33 crc kubenswrapper[4726]: I1123 20:35:33.046470 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-s7rvr"] Nov 23 20:35:34 crc kubenswrapper[4726]: I1123 20:35:34.629093 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97c02ff5-89f0-4834-ab15-1d226f632e17" path="/var/lib/kubelet/pods/97c02ff5-89f0-4834-ab15-1d226f632e17/volumes" Nov 23 20:35:37 crc kubenswrapper[4726]: I1123 20:35:37.343602 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:37 crc kubenswrapper[4726]: I1123 20:35:37.344149 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:37 crc kubenswrapper[4726]: I1123 20:35:37.403441 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:38 crc kubenswrapper[4726]: I1123 20:35:38.041282 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-fq2c6"] Nov 23 20:35:38 crc kubenswrapper[4726]: I1123 20:35:38.049707 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-fq2c6"] Nov 23 20:35:38 crc kubenswrapper[4726]: I1123 20:35:38.060539 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:38 crc kubenswrapper[4726]: I1123 20:35:38.112354 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ntj8n"] Nov 23 20:35:38 crc kubenswrapper[4726]: I1123 20:35:38.602184 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d6b028-7e7a-4c1d-a05e-5f49976b823d" path="/var/lib/kubelet/pods/b3d6b028-7e7a-4c1d-a05e-5f49976b823d/volumes" Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.005486 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ntj8n" podUID="30b465cd-2700-4659-9627-42833b738211" containerName="registry-server" containerID="cri-o://4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d" gracePeriod=2 Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.477598 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.533787 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cnq9\" (UniqueName: \"kubernetes.io/projected/30b465cd-2700-4659-9627-42833b738211-kube-api-access-2cnq9\") pod \"30b465cd-2700-4659-9627-42833b738211\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.533897 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-utilities\") pod \"30b465cd-2700-4659-9627-42833b738211\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.533922 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-catalog-content\") pod \"30b465cd-2700-4659-9627-42833b738211\" (UID: \"30b465cd-2700-4659-9627-42833b738211\") " Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.539162 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-utilities" (OuterVolumeSpecName: "utilities") pod "30b465cd-2700-4659-9627-42833b738211" (UID: "30b465cd-2700-4659-9627-42833b738211"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.558121 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b465cd-2700-4659-9627-42833b738211-kube-api-access-2cnq9" (OuterVolumeSpecName: "kube-api-access-2cnq9") pod "30b465cd-2700-4659-9627-42833b738211" (UID: "30b465cd-2700-4659-9627-42833b738211"). InnerVolumeSpecName "kube-api-access-2cnq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.595767 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30b465cd-2700-4659-9627-42833b738211" (UID: "30b465cd-2700-4659-9627-42833b738211"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.636566 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.636604 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30b465cd-2700-4659-9627-42833b738211-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:40 crc kubenswrapper[4726]: I1123 20:35:40.636620 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cnq9\" (UniqueName: \"kubernetes.io/projected/30b465cd-2700-4659-9627-42833b738211-kube-api-access-2cnq9\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.021398 4726 generic.go:334] "Generic (PLEG): container finished" podID="30b465cd-2700-4659-9627-42833b738211" containerID="4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d" exitCode=0 Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.021458 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ntj8n" event={"ID":"30b465cd-2700-4659-9627-42833b738211","Type":"ContainerDied","Data":"4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d"} Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.021501 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ntj8n" event={"ID":"30b465cd-2700-4659-9627-42833b738211","Type":"ContainerDied","Data":"f134aac5aadb35871132015405e41fc80a297d2a2353f9e04f305150b82a1d99"} Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.021534 4726 scope.go:117] "RemoveContainer" containerID="4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.021576 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ntj8n" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.049006 4726 scope.go:117] "RemoveContainer" containerID="87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.058731 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ntj8n"] Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.071701 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ntj8n"] Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.087541 4726 scope.go:117] "RemoveContainer" containerID="f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.147328 4726 scope.go:117] "RemoveContainer" containerID="4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d" Nov 23 20:35:41 crc kubenswrapper[4726]: E1123 20:35:41.150343 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d\": container with ID starting with 4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d not found: ID does not exist" containerID="4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.150403 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d"} err="failed to get container status \"4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d\": rpc error: code = NotFound desc = could not find container \"4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d\": container with ID starting with 4423af1b1e8264db493c13089f6966bbb42737b1dd2335dfc10e9d9fa5e7241d not found: ID does not exist" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.150450 4726 scope.go:117] "RemoveContainer" containerID="87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a" Nov 23 20:35:41 crc kubenswrapper[4726]: E1123 20:35:41.151077 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a\": container with ID starting with 87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a not found: ID does not exist" containerID="87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.151108 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a"} err="failed to get container status \"87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a\": rpc error: code = NotFound desc = could not find container \"87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a\": container with ID starting with 87f68420455b0e2b72ce54ff706fd2551b49e1aea4f5e5608faa92a3a107222a not found: ID does not exist" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.151129 4726 scope.go:117] "RemoveContainer" containerID="f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9" Nov 23 20:35:41 crc kubenswrapper[4726]: E1123 20:35:41.151538 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9\": container with ID starting with f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9 not found: ID does not exist" containerID="f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.151564 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9"} err="failed to get container status \"f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9\": rpc error: code = NotFound desc = could not find container \"f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9\": container with ID starting with f0b04ec56b88bb316718e7bcba107f02912cd0c6e0501ee696dea99d60e517c9 not found: ID does not exist" Nov 23 20:35:41 crc kubenswrapper[4726]: I1123 20:35:41.589135 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:35:41 crc kubenswrapper[4726]: E1123 20:35:41.589474 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:35:42 crc kubenswrapper[4726]: I1123 20:35:42.613032 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b465cd-2700-4659-9627-42833b738211" path="/var/lib/kubelet/pods/30b465cd-2700-4659-9627-42833b738211/volumes" Nov 23 20:35:56 crc kubenswrapper[4726]: I1123 20:35:56.168117 4726 generic.go:334] "Generic (PLEG): container finished" podID="70977db3-0e6c-4718-bebe-1d14ad030b9d" containerID="6279b72420a8ef5e3fd08686627f37edef6851aad393e3a83b76912dba0ce494" exitCode=0 Nov 23 20:35:56 crc kubenswrapper[4726]: I1123 20:35:56.168225 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" event={"ID":"70977db3-0e6c-4718-bebe-1d14ad030b9d","Type":"ContainerDied","Data":"6279b72420a8ef5e3fd08686627f37edef6851aad393e3a83b76912dba0ce494"} Nov 23 20:35:56 crc kubenswrapper[4726]: I1123 20:35:56.589005 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:35:56 crc kubenswrapper[4726]: E1123 20:35:56.589347 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.589217 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.746338 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-ssh-key\") pod \"70977db3-0e6c-4718-bebe-1d14ad030b9d\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.746634 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-inventory\") pod \"70977db3-0e6c-4718-bebe-1d14ad030b9d\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.746749 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftpv8\" (UniqueName: \"kubernetes.io/projected/70977db3-0e6c-4718-bebe-1d14ad030b9d-kube-api-access-ftpv8\") pod \"70977db3-0e6c-4718-bebe-1d14ad030b9d\" (UID: \"70977db3-0e6c-4718-bebe-1d14ad030b9d\") " Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.756180 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70977db3-0e6c-4718-bebe-1d14ad030b9d-kube-api-access-ftpv8" (OuterVolumeSpecName: "kube-api-access-ftpv8") pod "70977db3-0e6c-4718-bebe-1d14ad030b9d" (UID: "70977db3-0e6c-4718-bebe-1d14ad030b9d"). InnerVolumeSpecName "kube-api-access-ftpv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.773111 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-inventory" (OuterVolumeSpecName: "inventory") pod "70977db3-0e6c-4718-bebe-1d14ad030b9d" (UID: "70977db3-0e6c-4718-bebe-1d14ad030b9d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.797246 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "70977db3-0e6c-4718-bebe-1d14ad030b9d" (UID: "70977db3-0e6c-4718-bebe-1d14ad030b9d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.849734 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.849760 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftpv8\" (UniqueName: \"kubernetes.io/projected/70977db3-0e6c-4718-bebe-1d14ad030b9d-kube-api-access-ftpv8\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:57 crc kubenswrapper[4726]: I1123 20:35:57.849772 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70977db3-0e6c-4718-bebe-1d14ad030b9d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.188334 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" event={"ID":"70977db3-0e6c-4718-bebe-1d14ad030b9d","Type":"ContainerDied","Data":"6918c48b51dcb5807ff342bf93892bb757f7b3f73a56542cf69caec1ce7ab801"} Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.188561 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6918c48b51dcb5807ff342bf93892bb757f7b3f73a56542cf69caec1ce7ab801" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.188421 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.281205 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5"] Nov 23 20:35:58 crc kubenswrapper[4726]: E1123 20:35:58.281624 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b465cd-2700-4659-9627-42833b738211" containerName="extract-utilities" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.281645 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b465cd-2700-4659-9627-42833b738211" containerName="extract-utilities" Nov 23 20:35:58 crc kubenswrapper[4726]: E1123 20:35:58.281673 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b465cd-2700-4659-9627-42833b738211" containerName="registry-server" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.281683 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b465cd-2700-4659-9627-42833b738211" containerName="registry-server" Nov 23 20:35:58 crc kubenswrapper[4726]: E1123 20:35:58.281702 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70977db3-0e6c-4718-bebe-1d14ad030b9d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.281715 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="70977db3-0e6c-4718-bebe-1d14ad030b9d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:35:58 crc kubenswrapper[4726]: E1123 20:35:58.281767 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b465cd-2700-4659-9627-42833b738211" containerName="extract-content" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.281776 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b465cd-2700-4659-9627-42833b738211" containerName="extract-content" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.282055 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="70977db3-0e6c-4718-bebe-1d14ad030b9d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.282080 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b465cd-2700-4659-9627-42833b738211" containerName="registry-server" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.282778 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.285115 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.285311 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.286741 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.287946 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.303154 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5"] Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.460759 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.460985 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4skk\" (UniqueName: \"kubernetes.io/projected/e8139270-4ff9-440e-b998-aea1493900c4-kube-api-access-c4skk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.461036 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.562979 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.563093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4skk\" (UniqueName: \"kubernetes.io/projected/e8139270-4ff9-440e-b998-aea1493900c4-kube-api-access-c4skk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.563135 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.570280 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.574208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.586567 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4skk\" (UniqueName: \"kubernetes.io/projected/e8139270-4ff9-440e-b998-aea1493900c4-kube-api-access-c4skk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:58 crc kubenswrapper[4726]: I1123 20:35:58.611509 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:35:59 crc kubenswrapper[4726]: I1123 20:35:59.141165 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5"] Nov 23 20:35:59 crc kubenswrapper[4726]: W1123 20:35:59.143132 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8139270_4ff9_440e_b998_aea1493900c4.slice/crio-47d07fc997a3cb67d99fc543996c16143541035afdc719bd0e2dd95b95344668 WatchSource:0}: Error finding container 47d07fc997a3cb67d99fc543996c16143541035afdc719bd0e2dd95b95344668: Status 404 returned error can't find the container with id 47d07fc997a3cb67d99fc543996c16143541035afdc719bd0e2dd95b95344668 Nov 23 20:35:59 crc kubenswrapper[4726]: I1123 20:35:59.197097 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" event={"ID":"e8139270-4ff9-440e-b998-aea1493900c4","Type":"ContainerStarted","Data":"47d07fc997a3cb67d99fc543996c16143541035afdc719bd0e2dd95b95344668"} Nov 23 20:36:00 crc kubenswrapper[4726]: I1123 20:36:00.230509 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" event={"ID":"e8139270-4ff9-440e-b998-aea1493900c4","Type":"ContainerStarted","Data":"12999a5709031f4664e1f46234d94e7e243fb971f53e77b58ba815d48d285a98"} Nov 23 20:36:00 crc kubenswrapper[4726]: I1123 20:36:00.252144 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" podStartSLOduration=1.832805281 podStartE2EDuration="2.25212366s" podCreationTimestamp="2025-11-23 20:35:58 +0000 UTC" firstStartedPulling="2025-11-23 20:35:59.145026495 +0000 UTC m=+1667.294067451" lastFinishedPulling="2025-11-23 20:35:59.564344854 +0000 UTC m=+1667.713385830" observedRunningTime="2025-11-23 20:36:00.24612893 +0000 UTC m=+1668.395169896" watchObservedRunningTime="2025-11-23 20:36:00.25212366 +0000 UTC m=+1668.401164616" Nov 23 20:36:04 crc kubenswrapper[4726]: I1123 20:36:04.269845 4726 generic.go:334] "Generic (PLEG): container finished" podID="e8139270-4ff9-440e-b998-aea1493900c4" containerID="12999a5709031f4664e1f46234d94e7e243fb971f53e77b58ba815d48d285a98" exitCode=0 Nov 23 20:36:04 crc kubenswrapper[4726]: I1123 20:36:04.269931 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" event={"ID":"e8139270-4ff9-440e-b998-aea1493900c4","Type":"ContainerDied","Data":"12999a5709031f4664e1f46234d94e7e243fb971f53e77b58ba815d48d285a98"} Nov 23 20:36:05 crc kubenswrapper[4726]: I1123 20:36:05.782941 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:36:05 crc kubenswrapper[4726]: I1123 20:36:05.904121 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4skk\" (UniqueName: \"kubernetes.io/projected/e8139270-4ff9-440e-b998-aea1493900c4-kube-api-access-c4skk\") pod \"e8139270-4ff9-440e-b998-aea1493900c4\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " Nov 23 20:36:05 crc kubenswrapper[4726]: I1123 20:36:05.904296 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-inventory\") pod \"e8139270-4ff9-440e-b998-aea1493900c4\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " Nov 23 20:36:05 crc kubenswrapper[4726]: I1123 20:36:05.904435 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-ssh-key\") pod \"e8139270-4ff9-440e-b998-aea1493900c4\" (UID: \"e8139270-4ff9-440e-b998-aea1493900c4\") " Nov 23 20:36:05 crc kubenswrapper[4726]: I1123 20:36:05.911649 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8139270-4ff9-440e-b998-aea1493900c4-kube-api-access-c4skk" (OuterVolumeSpecName: "kube-api-access-c4skk") pod "e8139270-4ff9-440e-b998-aea1493900c4" (UID: "e8139270-4ff9-440e-b998-aea1493900c4"). InnerVolumeSpecName "kube-api-access-c4skk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:36:05 crc kubenswrapper[4726]: I1123 20:36:05.932883 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e8139270-4ff9-440e-b998-aea1493900c4" (UID: "e8139270-4ff9-440e-b998-aea1493900c4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:36:05 crc kubenswrapper[4726]: I1123 20:36:05.933554 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-inventory" (OuterVolumeSpecName: "inventory") pod "e8139270-4ff9-440e-b998-aea1493900c4" (UID: "e8139270-4ff9-440e-b998-aea1493900c4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.009335 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4skk\" (UniqueName: \"kubernetes.io/projected/e8139270-4ff9-440e-b998-aea1493900c4-kube-api-access-c4skk\") on node \"crc\" DevicePath \"\"" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.009820 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.010005 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e8139270-4ff9-440e-b998-aea1493900c4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.296440 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" event={"ID":"e8139270-4ff9-440e-b998-aea1493900c4","Type":"ContainerDied","Data":"47d07fc997a3cb67d99fc543996c16143541035afdc719bd0e2dd95b95344668"} Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.296525 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47d07fc997a3cb67d99fc543996c16143541035afdc719bd0e2dd95b95344668" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.296567 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.386253 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4"] Nov 23 20:36:06 crc kubenswrapper[4726]: E1123 20:36:06.387305 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8139270-4ff9-440e-b998-aea1493900c4" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.387335 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8139270-4ff9-440e-b998-aea1493900c4" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.387600 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8139270-4ff9-440e-b998-aea1493900c4" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.388513 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.392982 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.393262 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.394746 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.403990 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.416964 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4"] Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.520615 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.520714 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.520762 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cxrt\" (UniqueName: \"kubernetes.io/projected/c23e6d0b-0b2d-43cd-96f1-556499de1202-kube-api-access-9cxrt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.622522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.622639 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.622673 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cxrt\" (UniqueName: \"kubernetes.io/projected/c23e6d0b-0b2d-43cd-96f1-556499de1202-kube-api-access-9cxrt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.627513 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.629481 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.650317 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cxrt\" (UniqueName: \"kubernetes.io/projected/c23e6d0b-0b2d-43cd-96f1-556499de1202-kube-api-access-9cxrt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:06 crc kubenswrapper[4726]: I1123 20:36:06.707902 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:36:07 crc kubenswrapper[4726]: I1123 20:36:07.256177 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4"] Nov 23 20:36:07 crc kubenswrapper[4726]: W1123 20:36:07.274164 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc23e6d0b_0b2d_43cd_96f1_556499de1202.slice/crio-3a8daface486e4a5928dba34cd5986249a1795823261a1df65fb7fd6fdf06de5 WatchSource:0}: Error finding container 3a8daface486e4a5928dba34cd5986249a1795823261a1df65fb7fd6fdf06de5: Status 404 returned error can't find the container with id 3a8daface486e4a5928dba34cd5986249a1795823261a1df65fb7fd6fdf06de5 Nov 23 20:36:07 crc kubenswrapper[4726]: I1123 20:36:07.305575 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" event={"ID":"c23e6d0b-0b2d-43cd-96f1-556499de1202","Type":"ContainerStarted","Data":"3a8daface486e4a5928dba34cd5986249a1795823261a1df65fb7fd6fdf06de5"} Nov 23 20:36:08 crc kubenswrapper[4726]: I1123 20:36:08.318830 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" event={"ID":"c23e6d0b-0b2d-43cd-96f1-556499de1202","Type":"ContainerStarted","Data":"4f9241e8934dc5253b19735fd14ccfc3b88975a33e356abf2adf5d0d66a18e64"} Nov 23 20:36:08 crc kubenswrapper[4726]: I1123 20:36:08.349247 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" podStartSLOduration=1.873429515 podStartE2EDuration="2.349216424s" podCreationTimestamp="2025-11-23 20:36:06 +0000 UTC" firstStartedPulling="2025-11-23 20:36:07.275780743 +0000 UTC m=+1675.424821749" lastFinishedPulling="2025-11-23 20:36:07.751567672 +0000 UTC m=+1675.900608658" observedRunningTime="2025-11-23 20:36:08.338258603 +0000 UTC m=+1676.487299629" watchObservedRunningTime="2025-11-23 20:36:08.349216424 +0000 UTC m=+1676.498257410" Nov 23 20:36:10 crc kubenswrapper[4726]: I1123 20:36:10.591125 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:36:10 crc kubenswrapper[4726]: E1123 20:36:10.591721 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:36:18 crc kubenswrapper[4726]: I1123 20:36:18.211913 4726 scope.go:117] "RemoveContainer" containerID="156ea67dd89b7b3cbe69b38612407c2015921a70c503bdf55fcbfc3a2e8c8960" Nov 23 20:36:18 crc kubenswrapper[4726]: I1123 20:36:18.259285 4726 scope.go:117] "RemoveContainer" containerID="9e457b2e93890a3c0fe8b7a0d2d02dabbf8b6fe0371ca9884d190bb05668e590" Nov 23 20:36:18 crc kubenswrapper[4726]: I1123 20:36:18.299356 4726 scope.go:117] "RemoveContainer" containerID="603e083de46ca863689b365fdd8a22dfeba96b03369b30aa6e6bb0ca213e1e79" Nov 23 20:36:18 crc kubenswrapper[4726]: I1123 20:36:18.355249 4726 scope.go:117] "RemoveContainer" containerID="61c6c2fdc4e47dac3d875f5aaf8e00b45b2bb58a94ec35bed7f245267e41db62" Nov 23 20:36:18 crc kubenswrapper[4726]: I1123 20:36:18.385858 4726 scope.go:117] "RemoveContainer" containerID="d7a28f3138cbde6b4e68d7490eeef283ac368027bf675200afddf14efa2054ff" Nov 23 20:36:23 crc kubenswrapper[4726]: I1123 20:36:23.589655 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:36:23 crc kubenswrapper[4726]: E1123 20:36:23.590563 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.087447 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-45xfq"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.101272 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-186c-account-create-x2lmc"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.125402 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xbng7"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.133468 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-06fb-account-create-s7l7m"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.139643 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-bw8f5"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.153229 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8822-account-create-fxsgq"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.172306 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-45xfq"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.172606 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-bw8f5"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.183575 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8822-account-create-fxsgq"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.189835 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-186c-account-create-x2lmc"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.195829 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-06fb-account-create-s7l7m"] Nov 23 20:36:27 crc kubenswrapper[4726]: I1123 20:36:27.201562 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xbng7"] Nov 23 20:36:28 crc kubenswrapper[4726]: I1123 20:36:28.601596 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08af9ecf-3570-4e1c-98bd-163b5116634a" path="/var/lib/kubelet/pods/08af9ecf-3570-4e1c-98bd-163b5116634a/volumes" Nov 23 20:36:28 crc kubenswrapper[4726]: I1123 20:36:28.602469 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="100d7602-9a19-4155-b38b-f7bfd4b685ad" path="/var/lib/kubelet/pods/100d7602-9a19-4155-b38b-f7bfd4b685ad/volumes" Nov 23 20:36:28 crc kubenswrapper[4726]: I1123 20:36:28.603446 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3707a249-9a42-4ee2-8f07-c4a72c295a46" path="/var/lib/kubelet/pods/3707a249-9a42-4ee2-8f07-c4a72c295a46/volumes" Nov 23 20:36:28 crc kubenswrapper[4726]: I1123 20:36:28.604412 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79707a2d-f705-455f-a097-50c2e30eed2f" path="/var/lib/kubelet/pods/79707a2d-f705-455f-a097-50c2e30eed2f/volumes" Nov 23 20:36:28 crc kubenswrapper[4726]: I1123 20:36:28.606243 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2" path="/var/lib/kubelet/pods/a1b2b6e0-5c6e-4a35-95ae-84c61df4d3a2/volumes" Nov 23 20:36:28 crc kubenswrapper[4726]: I1123 20:36:28.607165 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb3fda45-1598-4496-9463-89d4f1cc2ef5" path="/var/lib/kubelet/pods/bb3fda45-1598-4496-9463-89d4f1cc2ef5/volumes" Nov 23 20:36:34 crc kubenswrapper[4726]: I1123 20:36:34.589507 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:36:34 crc kubenswrapper[4726]: E1123 20:36:34.591573 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:36:46 crc kubenswrapper[4726]: I1123 20:36:46.590052 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:36:46 crc kubenswrapper[4726]: E1123 20:36:46.591092 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:36:54 crc kubenswrapper[4726]: I1123 20:36:54.040467 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z4vqv"] Nov 23 20:36:54 crc kubenswrapper[4726]: I1123 20:36:54.046930 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z4vqv"] Nov 23 20:36:54 crc kubenswrapper[4726]: I1123 20:36:54.604259 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="094675fb-839b-47fe-b1cd-dec2fafbf906" path="/var/lib/kubelet/pods/094675fb-839b-47fe-b1cd-dec2fafbf906/volumes" Nov 23 20:36:59 crc kubenswrapper[4726]: I1123 20:36:59.588918 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:36:59 crc kubenswrapper[4726]: E1123 20:36:59.589823 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:37:10 crc kubenswrapper[4726]: I1123 20:37:10.940057 4726 generic.go:334] "Generic (PLEG): container finished" podID="c23e6d0b-0b2d-43cd-96f1-556499de1202" containerID="4f9241e8934dc5253b19735fd14ccfc3b88975a33e356abf2adf5d0d66a18e64" exitCode=0 Nov 23 20:37:10 crc kubenswrapper[4726]: I1123 20:37:10.940152 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" event={"ID":"c23e6d0b-0b2d-43cd-96f1-556499de1202","Type":"ContainerDied","Data":"4f9241e8934dc5253b19735fd14ccfc3b88975a33e356abf2adf5d0d66a18e64"} Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.465359 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.564051 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-ssh-key\") pod \"c23e6d0b-0b2d-43cd-96f1-556499de1202\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.564140 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cxrt\" (UniqueName: \"kubernetes.io/projected/c23e6d0b-0b2d-43cd-96f1-556499de1202-kube-api-access-9cxrt\") pod \"c23e6d0b-0b2d-43cd-96f1-556499de1202\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.564232 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-inventory\") pod \"c23e6d0b-0b2d-43cd-96f1-556499de1202\" (UID: \"c23e6d0b-0b2d-43cd-96f1-556499de1202\") " Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.582154 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c23e6d0b-0b2d-43cd-96f1-556499de1202-kube-api-access-9cxrt" (OuterVolumeSpecName: "kube-api-access-9cxrt") pod "c23e6d0b-0b2d-43cd-96f1-556499de1202" (UID: "c23e6d0b-0b2d-43cd-96f1-556499de1202"). InnerVolumeSpecName "kube-api-access-9cxrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.596490 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:37:12 crc kubenswrapper[4726]: E1123 20:37:12.596812 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.599776 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c23e6d0b-0b2d-43cd-96f1-556499de1202" (UID: "c23e6d0b-0b2d-43cd-96f1-556499de1202"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.600608 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-inventory" (OuterVolumeSpecName: "inventory") pod "c23e6d0b-0b2d-43cd-96f1-556499de1202" (UID: "c23e6d0b-0b2d-43cd-96f1-556499de1202"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.665836 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.666202 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cxrt\" (UniqueName: \"kubernetes.io/projected/c23e6d0b-0b2d-43cd-96f1-556499de1202-kube-api-access-9cxrt\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.666216 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c23e6d0b-0b2d-43cd-96f1-556499de1202-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.964787 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" event={"ID":"c23e6d0b-0b2d-43cd-96f1-556499de1202","Type":"ContainerDied","Data":"3a8daface486e4a5928dba34cd5986249a1795823261a1df65fb7fd6fdf06de5"} Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.964832 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a8daface486e4a5928dba34cd5986249a1795823261a1df65fb7fd6fdf06de5" Nov 23 20:37:12 crc kubenswrapper[4726]: I1123 20:37:12.964982 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.088491 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-srr4b"] Nov 23 20:37:13 crc kubenswrapper[4726]: E1123 20:37:13.088821 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23e6d0b-0b2d-43cd-96f1-556499de1202" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.088838 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23e6d0b-0b2d-43cd-96f1-556499de1202" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.089039 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c23e6d0b-0b2d-43cd-96f1-556499de1202" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.089626 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.092666 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.093091 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.093306 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.093491 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.109846 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-srr4b"] Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.177351 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.177396 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.177431 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjlkj\" (UniqueName: \"kubernetes.io/projected/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-kube-api-access-vjlkj\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.278451 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.278495 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.278529 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjlkj\" (UniqueName: \"kubernetes.io/projected/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-kube-api-access-vjlkj\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.282232 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.292832 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.293769 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjlkj\" (UniqueName: \"kubernetes.io/projected/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-kube-api-access-vjlkj\") pod \"ssh-known-hosts-edpm-deployment-srr4b\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.404079 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.953559 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-srr4b"] Nov 23 20:37:13 crc kubenswrapper[4726]: I1123 20:37:13.992231 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" event={"ID":"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6","Type":"ContainerStarted","Data":"a9d4685d42ad26ac1cc8a06ff994ffd2cf106637560fd99b38bc15033c669118"} Nov 23 20:37:15 crc kubenswrapper[4726]: I1123 20:37:15.004666 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" event={"ID":"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6","Type":"ContainerStarted","Data":"5b131b0e71d29f3c72f90ecf13a46837967d4af91e9c038293dbfc781f195501"} Nov 23 20:37:15 crc kubenswrapper[4726]: I1123 20:37:15.026949 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" podStartSLOduration=1.5494356969999998 podStartE2EDuration="2.026931749s" podCreationTimestamp="2025-11-23 20:37:13 +0000 UTC" firstStartedPulling="2025-11-23 20:37:13.978664745 +0000 UTC m=+1742.127705711" lastFinishedPulling="2025-11-23 20:37:14.456160797 +0000 UTC m=+1742.605201763" observedRunningTime="2025-11-23 20:37:15.022758267 +0000 UTC m=+1743.171799243" watchObservedRunningTime="2025-11-23 20:37:15.026931749 +0000 UTC m=+1743.175972705" Nov 23 20:37:18 crc kubenswrapper[4726]: I1123 20:37:18.548027 4726 scope.go:117] "RemoveContainer" containerID="ceebb89ddb0e75b4260a3d34ccc2eedc24c0febca6e177d82bda7d26ac1722a8" Nov 23 20:37:18 crc kubenswrapper[4726]: I1123 20:37:18.579916 4726 scope.go:117] "RemoveContainer" containerID="05f0b309c0be90990868e64f86abfdda00d40fe6b71cd616e2151dd989f0bde5" Nov 23 20:37:18 crc kubenswrapper[4726]: I1123 20:37:18.621552 4726 scope.go:117] "RemoveContainer" containerID="cd4da0aef4c26bd531f7b1d20e1f587bf7ba4109c52ea55ae21d0296fc521d6f" Nov 23 20:37:18 crc kubenswrapper[4726]: I1123 20:37:18.656474 4726 scope.go:117] "RemoveContainer" containerID="5882b6fe29c83f14108e1f4ee991c9038b632d758d4db3ca9b62ef0cf8deafa4" Nov 23 20:37:18 crc kubenswrapper[4726]: I1123 20:37:18.723791 4726 scope.go:117] "RemoveContainer" containerID="0774a1b0a35dd1676ffe601d3b108f18550694efa16a719335cd55dfca76d765" Nov 23 20:37:18 crc kubenswrapper[4726]: I1123 20:37:18.743566 4726 scope.go:117] "RemoveContainer" containerID="1dab6608897606fd5c259c5c7de8625df3c33a89e7745f46c17a4b89fed137d6" Nov 23 20:37:18 crc kubenswrapper[4726]: I1123 20:37:18.779217 4726 scope.go:117] "RemoveContainer" containerID="aa5df2feac6a8cc66913a2834fbc6eb25d7d8a29c868d7b1de754e0a20640773" Nov 23 20:37:23 crc kubenswrapper[4726]: I1123 20:37:23.063568 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qps7j"] Nov 23 20:37:23 crc kubenswrapper[4726]: I1123 20:37:23.078287 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-dqb9v"] Nov 23 20:37:23 crc kubenswrapper[4726]: I1123 20:37:23.083181 4726 generic.go:334] "Generic (PLEG): container finished" podID="6c2cb05b-a398-4e68-b9e3-b660e47ce4d6" containerID="5b131b0e71d29f3c72f90ecf13a46837967d4af91e9c038293dbfc781f195501" exitCode=0 Nov 23 20:37:23 crc kubenswrapper[4726]: I1123 20:37:23.083207 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" event={"ID":"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6","Type":"ContainerDied","Data":"5b131b0e71d29f3c72f90ecf13a46837967d4af91e9c038293dbfc781f195501"} Nov 23 20:37:23 crc kubenswrapper[4726]: I1123 20:37:23.086999 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qps7j"] Nov 23 20:37:23 crc kubenswrapper[4726]: I1123 20:37:23.099357 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-dqb9v"] Nov 23 20:37:23 crc kubenswrapper[4726]: I1123 20:37:23.592671 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:37:23 crc kubenswrapper[4726]: E1123 20:37:23.593192 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.451527 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.603700 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80fe916a-801a-4d1e-a9b1-fb819406e39f" path="/var/lib/kubelet/pods/80fe916a-801a-4d1e-a9b1-fb819406e39f/volumes" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.604558 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a8bd6c7-c313-475a-95ce-d95fb659ebaf" path="/var/lib/kubelet/pods/8a8bd6c7-c313-475a-95ce-d95fb659ebaf/volumes" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.607985 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjlkj\" (UniqueName: \"kubernetes.io/projected/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-kube-api-access-vjlkj\") pod \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.608069 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-inventory-0\") pod \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.608176 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-ssh-key-openstack-edpm-ipam\") pod \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\" (UID: \"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6\") " Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.614508 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-kube-api-access-vjlkj" (OuterVolumeSpecName: "kube-api-access-vjlkj") pod "6c2cb05b-a398-4e68-b9e3-b660e47ce4d6" (UID: "6c2cb05b-a398-4e68-b9e3-b660e47ce4d6"). InnerVolumeSpecName "kube-api-access-vjlkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.641818 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6c2cb05b-a398-4e68-b9e3-b660e47ce4d6" (UID: "6c2cb05b-a398-4e68-b9e3-b660e47ce4d6"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.648173 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6c2cb05b-a398-4e68-b9e3-b660e47ce4d6" (UID: "6c2cb05b-a398-4e68-b9e3-b660e47ce4d6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.710512 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjlkj\" (UniqueName: \"kubernetes.io/projected/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-kube-api-access-vjlkj\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.710680 4726 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:24 crc kubenswrapper[4726]: I1123 20:37:24.710753 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.101164 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" event={"ID":"6c2cb05b-a398-4e68-b9e3-b660e47ce4d6","Type":"ContainerDied","Data":"a9d4685d42ad26ac1cc8a06ff994ffd2cf106637560fd99b38bc15033c669118"} Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.101520 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9d4685d42ad26ac1cc8a06ff994ffd2cf106637560fd99b38bc15033c669118" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.101253 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-srr4b" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.196039 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw"] Nov 23 20:37:25 crc kubenswrapper[4726]: E1123 20:37:25.196584 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c2cb05b-a398-4e68-b9e3-b660e47ce4d6" containerName="ssh-known-hosts-edpm-deployment" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.196619 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c2cb05b-a398-4e68-b9e3-b660e47ce4d6" containerName="ssh-known-hosts-edpm-deployment" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.196986 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c2cb05b-a398-4e68-b9e3-b660e47ce4d6" containerName="ssh-known-hosts-edpm-deployment" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.197978 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.204826 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.205381 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.205576 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.205994 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.212194 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw"] Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.225415 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.225469 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.225578 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqfbm\" (UniqueName: \"kubernetes.io/projected/40ad6dc8-1508-4186-bbae-a5ff96044133-kube-api-access-hqfbm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.326967 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.327266 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.327483 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqfbm\" (UniqueName: \"kubernetes.io/projected/40ad6dc8-1508-4186-bbae-a5ff96044133-kube-api-access-hqfbm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.337694 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.337717 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.349475 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqfbm\" (UniqueName: \"kubernetes.io/projected/40ad6dc8-1508-4186-bbae-a5ff96044133-kube-api-access-hqfbm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9vvmw\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:25 crc kubenswrapper[4726]: I1123 20:37:25.518775 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:26 crc kubenswrapper[4726]: I1123 20:37:26.087973 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw"] Nov 23 20:37:26 crc kubenswrapper[4726]: I1123 20:37:26.115633 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" event={"ID":"40ad6dc8-1508-4186-bbae-a5ff96044133","Type":"ContainerStarted","Data":"4f871afd7061ff642ff806c33409f7ab205ece4d1e433ad282b12161049e797f"} Nov 23 20:37:27 crc kubenswrapper[4726]: I1123 20:37:27.134448 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" event={"ID":"40ad6dc8-1508-4186-bbae-a5ff96044133","Type":"ContainerStarted","Data":"655e80a4c0e6045c113f4cf5bca2964a39a0cf2a5fdd8f7fcfbe6fa26abb58c1"} Nov 23 20:37:27 crc kubenswrapper[4726]: I1123 20:37:27.170581 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" podStartSLOduration=1.7015539579999999 podStartE2EDuration="2.170559591s" podCreationTimestamp="2025-11-23 20:37:25 +0000 UTC" firstStartedPulling="2025-11-23 20:37:26.10201247 +0000 UTC m=+1754.251053436" lastFinishedPulling="2025-11-23 20:37:26.571018103 +0000 UTC m=+1754.720059069" observedRunningTime="2025-11-23 20:37:27.164124732 +0000 UTC m=+1755.313165688" watchObservedRunningTime="2025-11-23 20:37:27.170559591 +0000 UTC m=+1755.319600547" Nov 23 20:37:36 crc kubenswrapper[4726]: I1123 20:37:36.243097 4726 generic.go:334] "Generic (PLEG): container finished" podID="40ad6dc8-1508-4186-bbae-a5ff96044133" containerID="655e80a4c0e6045c113f4cf5bca2964a39a0cf2a5fdd8f7fcfbe6fa26abb58c1" exitCode=0 Nov 23 20:37:36 crc kubenswrapper[4726]: I1123 20:37:36.243735 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" event={"ID":"40ad6dc8-1508-4186-bbae-a5ff96044133","Type":"ContainerDied","Data":"655e80a4c0e6045c113f4cf5bca2964a39a0cf2a5fdd8f7fcfbe6fa26abb58c1"} Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.589358 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:37:37 crc kubenswrapper[4726]: E1123 20:37:37.590351 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.730247 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.879983 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqfbm\" (UniqueName: \"kubernetes.io/projected/40ad6dc8-1508-4186-bbae-a5ff96044133-kube-api-access-hqfbm\") pod \"40ad6dc8-1508-4186-bbae-a5ff96044133\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.880159 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-ssh-key\") pod \"40ad6dc8-1508-4186-bbae-a5ff96044133\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.880259 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-inventory\") pod \"40ad6dc8-1508-4186-bbae-a5ff96044133\" (UID: \"40ad6dc8-1508-4186-bbae-a5ff96044133\") " Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.885883 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40ad6dc8-1508-4186-bbae-a5ff96044133-kube-api-access-hqfbm" (OuterVolumeSpecName: "kube-api-access-hqfbm") pod "40ad6dc8-1508-4186-bbae-a5ff96044133" (UID: "40ad6dc8-1508-4186-bbae-a5ff96044133"). InnerVolumeSpecName "kube-api-access-hqfbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.910029 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-inventory" (OuterVolumeSpecName: "inventory") pod "40ad6dc8-1508-4186-bbae-a5ff96044133" (UID: "40ad6dc8-1508-4186-bbae-a5ff96044133"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.928133 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "40ad6dc8-1508-4186-bbae-a5ff96044133" (UID: "40ad6dc8-1508-4186-bbae-a5ff96044133"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.982368 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.982399 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqfbm\" (UniqueName: \"kubernetes.io/projected/40ad6dc8-1508-4186-bbae-a5ff96044133-kube-api-access-hqfbm\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:37 crc kubenswrapper[4726]: I1123 20:37:37.982410 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40ad6dc8-1508-4186-bbae-a5ff96044133-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.271730 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" event={"ID":"40ad6dc8-1508-4186-bbae-a5ff96044133","Type":"ContainerDied","Data":"4f871afd7061ff642ff806c33409f7ab205ece4d1e433ad282b12161049e797f"} Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.272322 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f871afd7061ff642ff806c33409f7ab205ece4d1e433ad282b12161049e797f" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.272406 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.350684 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz"] Nov 23 20:37:38 crc kubenswrapper[4726]: E1123 20:37:38.351164 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ad6dc8-1508-4186-bbae-a5ff96044133" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.351190 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ad6dc8-1508-4186-bbae-a5ff96044133" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.351463 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="40ad6dc8-1508-4186-bbae-a5ff96044133" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.352238 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.357743 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.359235 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.359379 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.359967 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.383558 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz"] Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.499666 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.499721 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29m8k\" (UniqueName: \"kubernetes.io/projected/2ad4f10a-15ed-4107-9c4e-16975854883e-kube-api-access-29m8k\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.499889 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.601324 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.601388 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29m8k\" (UniqueName: \"kubernetes.io/projected/2ad4f10a-15ed-4107-9c4e-16975854883e-kube-api-access-29m8k\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.601444 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.613058 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.629948 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.672439 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29m8k\" (UniqueName: \"kubernetes.io/projected/2ad4f10a-15ed-4107-9c4e-16975854883e-kube-api-access-29m8k\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:38 crc kubenswrapper[4726]: I1123 20:37:38.968223 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:39 crc kubenswrapper[4726]: I1123 20:37:39.493675 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz"] Nov 23 20:37:39 crc kubenswrapper[4726]: W1123 20:37:39.498403 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ad4f10a_15ed_4107_9c4e_16975854883e.slice/crio-4eef98d8a195364c7a2c40a153ca9d617ef0da3d23bd41f10be97ed840c2e22a WatchSource:0}: Error finding container 4eef98d8a195364c7a2c40a153ca9d617ef0da3d23bd41f10be97ed840c2e22a: Status 404 returned error can't find the container with id 4eef98d8a195364c7a2c40a153ca9d617ef0da3d23bd41f10be97ed840c2e22a Nov 23 20:37:40 crc kubenswrapper[4726]: I1123 20:37:40.289548 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" event={"ID":"2ad4f10a-15ed-4107-9c4e-16975854883e","Type":"ContainerStarted","Data":"ae51844d8d85283811611f7344fbf3bac3c0353ff46cc259a81987ca4964d54f"} Nov 23 20:37:40 crc kubenswrapper[4726]: I1123 20:37:40.290214 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" event={"ID":"2ad4f10a-15ed-4107-9c4e-16975854883e","Type":"ContainerStarted","Data":"4eef98d8a195364c7a2c40a153ca9d617ef0da3d23bd41f10be97ed840c2e22a"} Nov 23 20:37:40 crc kubenswrapper[4726]: I1123 20:37:40.318614 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" podStartSLOduration=1.92065151 podStartE2EDuration="2.318597947s" podCreationTimestamp="2025-11-23 20:37:38 +0000 UTC" firstStartedPulling="2025-11-23 20:37:39.500574164 +0000 UTC m=+1767.649615110" lastFinishedPulling="2025-11-23 20:37:39.898520581 +0000 UTC m=+1768.047561547" observedRunningTime="2025-11-23 20:37:40.309195601 +0000 UTC m=+1768.458236567" watchObservedRunningTime="2025-11-23 20:37:40.318597947 +0000 UTC m=+1768.467638903" Nov 23 20:37:49 crc kubenswrapper[4726]: I1123 20:37:49.588620 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:37:49 crc kubenswrapper[4726]: E1123 20:37:49.590324 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:37:50 crc kubenswrapper[4726]: I1123 20:37:50.394948 4726 generic.go:334] "Generic (PLEG): container finished" podID="2ad4f10a-15ed-4107-9c4e-16975854883e" containerID="ae51844d8d85283811611f7344fbf3bac3c0353ff46cc259a81987ca4964d54f" exitCode=0 Nov 23 20:37:50 crc kubenswrapper[4726]: I1123 20:37:50.394961 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" event={"ID":"2ad4f10a-15ed-4107-9c4e-16975854883e","Type":"ContainerDied","Data":"ae51844d8d85283811611f7344fbf3bac3c0353ff46cc259a81987ca4964d54f"} Nov 23 20:37:51 crc kubenswrapper[4726]: I1123 20:37:51.873604 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:37:51 crc kubenswrapper[4726]: I1123 20:37:51.963962 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29m8k\" (UniqueName: \"kubernetes.io/projected/2ad4f10a-15ed-4107-9c4e-16975854883e-kube-api-access-29m8k\") pod \"2ad4f10a-15ed-4107-9c4e-16975854883e\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " Nov 23 20:37:51 crc kubenswrapper[4726]: I1123 20:37:51.964131 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-inventory\") pod \"2ad4f10a-15ed-4107-9c4e-16975854883e\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " Nov 23 20:37:51 crc kubenswrapper[4726]: I1123 20:37:51.964177 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-ssh-key\") pod \"2ad4f10a-15ed-4107-9c4e-16975854883e\" (UID: \"2ad4f10a-15ed-4107-9c4e-16975854883e\") " Nov 23 20:37:51 crc kubenswrapper[4726]: I1123 20:37:51.972158 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ad4f10a-15ed-4107-9c4e-16975854883e-kube-api-access-29m8k" (OuterVolumeSpecName: "kube-api-access-29m8k") pod "2ad4f10a-15ed-4107-9c4e-16975854883e" (UID: "2ad4f10a-15ed-4107-9c4e-16975854883e"). InnerVolumeSpecName "kube-api-access-29m8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:37:51 crc kubenswrapper[4726]: I1123 20:37:51.993016 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-inventory" (OuterVolumeSpecName: "inventory") pod "2ad4f10a-15ed-4107-9c4e-16975854883e" (UID: "2ad4f10a-15ed-4107-9c4e-16975854883e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:37:51 crc kubenswrapper[4726]: I1123 20:37:51.993453 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ad4f10a-15ed-4107-9c4e-16975854883e" (UID: "2ad4f10a-15ed-4107-9c4e-16975854883e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:37:52 crc kubenswrapper[4726]: I1123 20:37:52.067581 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29m8k\" (UniqueName: \"kubernetes.io/projected/2ad4f10a-15ed-4107-9c4e-16975854883e-kube-api-access-29m8k\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:52 crc kubenswrapper[4726]: I1123 20:37:52.067619 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:52 crc kubenswrapper[4726]: I1123 20:37:52.067632 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ad4f10a-15ed-4107-9c4e-16975854883e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:37:52 crc kubenswrapper[4726]: I1123 20:37:52.420562 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" event={"ID":"2ad4f10a-15ed-4107-9c4e-16975854883e","Type":"ContainerDied","Data":"4eef98d8a195364c7a2c40a153ca9d617ef0da3d23bd41f10be97ed840c2e22a"} Nov 23 20:37:52 crc kubenswrapper[4726]: I1123 20:37:52.420602 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eef98d8a195364c7a2c40a153ca9d617ef0da3d23bd41f10be97ed840c2e22a" Nov 23 20:37:52 crc kubenswrapper[4726]: I1123 20:37:52.420646 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz" Nov 23 20:38:01 crc kubenswrapper[4726]: I1123 20:38:01.589107 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:38:01 crc kubenswrapper[4726]: E1123 20:38:01.590988 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:38:04 crc kubenswrapper[4726]: I1123 20:38:04.042147 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-8qvxw"] Nov 23 20:38:04 crc kubenswrapper[4726]: I1123 20:38:04.049719 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-8qvxw"] Nov 23 20:38:04 crc kubenswrapper[4726]: I1123 20:38:04.616748 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="267f6c6b-3b5d-4e89-bf80-3650b890aed9" path="/var/lib/kubelet/pods/267f6c6b-3b5d-4e89-bf80-3650b890aed9/volumes" Nov 23 20:38:12 crc kubenswrapper[4726]: I1123 20:38:12.600336 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:38:12 crc kubenswrapper[4726]: E1123 20:38:12.601264 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:38:18 crc kubenswrapper[4726]: I1123 20:38:18.893531 4726 scope.go:117] "RemoveContainer" containerID="e085e9ab6c370835a2062856c60669b7989070212f1dacc717ed01fb1746eace" Nov 23 20:38:19 crc kubenswrapper[4726]: I1123 20:38:19.007314 4726 scope.go:117] "RemoveContainer" containerID="cff0b94346727804ad9dcbf63d396db0137c49417b03f512ba5091ab5bb59225" Nov 23 20:38:19 crc kubenswrapper[4726]: I1123 20:38:19.060706 4726 scope.go:117] "RemoveContainer" containerID="f49aefbb1bd6feb8f32dec14ca09ac31201d7769cec209ae7c463464c7edf83b" Nov 23 20:38:24 crc kubenswrapper[4726]: I1123 20:38:24.590355 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:38:24 crc kubenswrapper[4726]: E1123 20:38:24.591630 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:38:39 crc kubenswrapper[4726]: I1123 20:38:39.589363 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:38:39 crc kubenswrapper[4726]: I1123 20:38:39.903996 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"b4944e760d4a10e30b574916d9b12b8294572948eb1263ca192e2958218c0945"} Nov 23 20:40:39 crc kubenswrapper[4726]: I1123 20:40:39.043233 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:40:39 crc kubenswrapper[4726]: I1123 20:40:39.043984 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:41:09 crc kubenswrapper[4726]: I1123 20:41:09.043327 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:41:09 crc kubenswrapper[4726]: I1123 20:41:09.043959 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.043219 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.043800 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.043861 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.044656 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b4944e760d4a10e30b574916d9b12b8294572948eb1263ca192e2958218c0945"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.044707 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://b4944e760d4a10e30b574916d9b12b8294572948eb1263ca192e2958218c0945" gracePeriod=600 Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.814010 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="b4944e760d4a10e30b574916d9b12b8294572948eb1263ca192e2958218c0945" exitCode=0 Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.814082 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"b4944e760d4a10e30b574916d9b12b8294572948eb1263ca192e2958218c0945"} Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.814735 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e"} Nov 23 20:41:39 crc kubenswrapper[4726]: I1123 20:41:39.814771 4726 scope.go:117] "RemoveContainer" containerID="28b24e799030d3fdde0d498d6f8d865f5d9858bbf84903eea30d15329cbf9cfa" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.695855 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kh9r6"] Nov 23 20:42:42 crc kubenswrapper[4726]: E1123 20:42:42.697201 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad4f10a-15ed-4107-9c4e-16975854883e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.697228 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad4f10a-15ed-4107-9c4e-16975854883e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.697732 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad4f10a-15ed-4107-9c4e-16975854883e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.705926 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.714619 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kh9r6"] Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.865581 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t4zv\" (UniqueName: \"kubernetes.io/projected/f8492bea-d49a-4832-abf4-9964651479cc-kube-api-access-8t4zv\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.866323 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-utilities\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.866349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-catalog-content\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.973527 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-utilities\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.973586 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-catalog-content\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.973782 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t4zv\" (UniqueName: \"kubernetes.io/projected/f8492bea-d49a-4832-abf4-9964651479cc-kube-api-access-8t4zv\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.974046 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-utilities\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:42 crc kubenswrapper[4726]: I1123 20:42:42.974062 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-catalog-content\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:43 crc kubenswrapper[4726]: I1123 20:42:43.005565 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t4zv\" (UniqueName: \"kubernetes.io/projected/f8492bea-d49a-4832-abf4-9964651479cc-kube-api-access-8t4zv\") pod \"redhat-operators-kh9r6\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:43 crc kubenswrapper[4726]: I1123 20:42:43.040816 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:43 crc kubenswrapper[4726]: I1123 20:42:43.523711 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kh9r6"] Nov 23 20:42:43 crc kubenswrapper[4726]: E1123 20:42:43.878523 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8492bea_d49a_4832_abf4_9964651479cc.slice/crio-a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8492bea_d49a_4832_abf4_9964651479cc.slice/crio-conmon-a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17.scope\": RecentStats: unable to find data in memory cache]" Nov 23 20:42:44 crc kubenswrapper[4726]: I1123 20:42:44.460153 4726 generic.go:334] "Generic (PLEG): container finished" podID="f8492bea-d49a-4832-abf4-9964651479cc" containerID="a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17" exitCode=0 Nov 23 20:42:44 crc kubenswrapper[4726]: I1123 20:42:44.460291 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh9r6" event={"ID":"f8492bea-d49a-4832-abf4-9964651479cc","Type":"ContainerDied","Data":"a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17"} Nov 23 20:42:44 crc kubenswrapper[4726]: I1123 20:42:44.462475 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh9r6" event={"ID":"f8492bea-d49a-4832-abf4-9964651479cc","Type":"ContainerStarted","Data":"5344c909f3992523e105160e752d38772c7221052a75a6d18c8cd9ec780635e3"} Nov 23 20:42:44 crc kubenswrapper[4726]: I1123 20:42:44.463393 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 20:42:45 crc kubenswrapper[4726]: I1123 20:42:45.469884 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh9r6" event={"ID":"f8492bea-d49a-4832-abf4-9964651479cc","Type":"ContainerStarted","Data":"1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48"} Nov 23 20:42:49 crc kubenswrapper[4726]: I1123 20:42:49.505082 4726 generic.go:334] "Generic (PLEG): container finished" podID="f8492bea-d49a-4832-abf4-9964651479cc" containerID="1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48" exitCode=0 Nov 23 20:42:49 crc kubenswrapper[4726]: I1123 20:42:49.505130 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh9r6" event={"ID":"f8492bea-d49a-4832-abf4-9964651479cc","Type":"ContainerDied","Data":"1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48"} Nov 23 20:42:50 crc kubenswrapper[4726]: I1123 20:42:50.516861 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh9r6" event={"ID":"f8492bea-d49a-4832-abf4-9964651479cc","Type":"ContainerStarted","Data":"2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3"} Nov 23 20:42:50 crc kubenswrapper[4726]: I1123 20:42:50.546597 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kh9r6" podStartSLOduration=2.944193066 podStartE2EDuration="8.546577109s" podCreationTimestamp="2025-11-23 20:42:42 +0000 UTC" firstStartedPulling="2025-11-23 20:42:44.463172061 +0000 UTC m=+2072.612213017" lastFinishedPulling="2025-11-23 20:42:50.065556084 +0000 UTC m=+2078.214597060" observedRunningTime="2025-11-23 20:42:50.536984858 +0000 UTC m=+2078.686025844" watchObservedRunningTime="2025-11-23 20:42:50.546577109 +0000 UTC m=+2078.695618075" Nov 23 20:42:53 crc kubenswrapper[4726]: I1123 20:42:53.041828 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:53 crc kubenswrapper[4726]: I1123 20:42:53.041890 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:42:54 crc kubenswrapper[4726]: I1123 20:42:54.102282 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kh9r6" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="registry-server" probeResult="failure" output=< Nov 23 20:42:54 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 20:42:54 crc kubenswrapper[4726]: > Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.485923 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9qvrg"] Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.488833 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.498354 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qvrg"] Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.623980 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-utilities\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.624262 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-catalog-content\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.624340 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgf58\" (UniqueName: \"kubernetes.io/projected/a4f03f8d-fa97-450c-9117-fd601115c842-kube-api-access-bgf58\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.725553 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgf58\" (UniqueName: \"kubernetes.io/projected/a4f03f8d-fa97-450c-9117-fd601115c842-kube-api-access-bgf58\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.725693 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-utilities\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.725728 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-catalog-content\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.726530 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-utilities\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.726537 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-catalog-content\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.750066 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgf58\" (UniqueName: \"kubernetes.io/projected/a4f03f8d-fa97-450c-9117-fd601115c842-kube-api-access-bgf58\") pod \"redhat-marketplace-9qvrg\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:55 crc kubenswrapper[4726]: I1123 20:42:55.846029 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:42:56 crc kubenswrapper[4726]: I1123 20:42:56.287554 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qvrg"] Nov 23 20:42:56 crc kubenswrapper[4726]: I1123 20:42:56.584257 4726 generic.go:334] "Generic (PLEG): container finished" podID="a4f03f8d-fa97-450c-9117-fd601115c842" containerID="84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e" exitCode=0 Nov 23 20:42:56 crc kubenswrapper[4726]: I1123 20:42:56.584473 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qvrg" event={"ID":"a4f03f8d-fa97-450c-9117-fd601115c842","Type":"ContainerDied","Data":"84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e"} Nov 23 20:42:56 crc kubenswrapper[4726]: I1123 20:42:56.584594 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qvrg" event={"ID":"a4f03f8d-fa97-450c-9117-fd601115c842","Type":"ContainerStarted","Data":"054436f058ab6dd72eaa44ef7e313905b40a810e868fb177d38cd197b46a4955"} Nov 23 20:42:57 crc kubenswrapper[4726]: I1123 20:42:57.596252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qvrg" event={"ID":"a4f03f8d-fa97-450c-9117-fd601115c842","Type":"ContainerStarted","Data":"b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd"} Nov 23 20:42:58 crc kubenswrapper[4726]: I1123 20:42:58.610361 4726 generic.go:334] "Generic (PLEG): container finished" podID="a4f03f8d-fa97-450c-9117-fd601115c842" containerID="b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd" exitCode=0 Nov 23 20:42:58 crc kubenswrapper[4726]: I1123 20:42:58.610402 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qvrg" event={"ID":"a4f03f8d-fa97-450c-9117-fd601115c842","Type":"ContainerDied","Data":"b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd"} Nov 23 20:42:59 crc kubenswrapper[4726]: I1123 20:42:59.619823 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qvrg" event={"ID":"a4f03f8d-fa97-450c-9117-fd601115c842","Type":"ContainerStarted","Data":"ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d"} Nov 23 20:42:59 crc kubenswrapper[4726]: I1123 20:42:59.649032 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9qvrg" podStartSLOduration=2.174928429 podStartE2EDuration="4.649014486s" podCreationTimestamp="2025-11-23 20:42:55 +0000 UTC" firstStartedPulling="2025-11-23 20:42:56.586267968 +0000 UTC m=+2084.735308914" lastFinishedPulling="2025-11-23 20:42:59.060354005 +0000 UTC m=+2087.209394971" observedRunningTime="2025-11-23 20:42:59.639311942 +0000 UTC m=+2087.788352898" watchObservedRunningTime="2025-11-23 20:42:59.649014486 +0000 UTC m=+2087.798055442" Nov 23 20:43:04 crc kubenswrapper[4726]: I1123 20:43:04.086259 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kh9r6" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="registry-server" probeResult="failure" output=< Nov 23 20:43:04 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 20:43:04 crc kubenswrapper[4726]: > Nov 23 20:43:05 crc kubenswrapper[4726]: I1123 20:43:05.846687 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:43:05 crc kubenswrapper[4726]: I1123 20:43:05.847469 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:43:05 crc kubenswrapper[4726]: I1123 20:43:05.921902 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:43:06 crc kubenswrapper[4726]: I1123 20:43:06.755309 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:43:06 crc kubenswrapper[4726]: I1123 20:43:06.807919 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qvrg"] Nov 23 20:43:08 crc kubenswrapper[4726]: I1123 20:43:08.716991 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9qvrg" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" containerName="registry-server" containerID="cri-o://ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d" gracePeriod=2 Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.557462 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.732314 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-catalog-content\") pod \"a4f03f8d-fa97-450c-9117-fd601115c842\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.732421 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-utilities\") pod \"a4f03f8d-fa97-450c-9117-fd601115c842\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.732558 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgf58\" (UniqueName: \"kubernetes.io/projected/a4f03f8d-fa97-450c-9117-fd601115c842-kube-api-access-bgf58\") pod \"a4f03f8d-fa97-450c-9117-fd601115c842\" (UID: \"a4f03f8d-fa97-450c-9117-fd601115c842\") " Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.733601 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-utilities" (OuterVolumeSpecName: "utilities") pod "a4f03f8d-fa97-450c-9117-fd601115c842" (UID: "a4f03f8d-fa97-450c-9117-fd601115c842"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.735469 4726 generic.go:334] "Generic (PLEG): container finished" podID="a4f03f8d-fa97-450c-9117-fd601115c842" containerID="ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d" exitCode=0 Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.735511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qvrg" event={"ID":"a4f03f8d-fa97-450c-9117-fd601115c842","Type":"ContainerDied","Data":"ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d"} Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.735523 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9qvrg" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.735550 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9qvrg" event={"ID":"a4f03f8d-fa97-450c-9117-fd601115c842","Type":"ContainerDied","Data":"054436f058ab6dd72eaa44ef7e313905b40a810e868fb177d38cd197b46a4955"} Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.735567 4726 scope.go:117] "RemoveContainer" containerID="ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.748743 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4f03f8d-fa97-450c-9117-fd601115c842-kube-api-access-bgf58" (OuterVolumeSpecName: "kube-api-access-bgf58") pod "a4f03f8d-fa97-450c-9117-fd601115c842" (UID: "a4f03f8d-fa97-450c-9117-fd601115c842"). InnerVolumeSpecName "kube-api-access-bgf58". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.755460 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4f03f8d-fa97-450c-9117-fd601115c842" (UID: "a4f03f8d-fa97-450c-9117-fd601115c842"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.793444 4726 scope.go:117] "RemoveContainer" containerID="b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.812373 4726 scope.go:117] "RemoveContainer" containerID="84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.834355 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.834387 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4f03f8d-fa97-450c-9117-fd601115c842-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.834402 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgf58\" (UniqueName: \"kubernetes.io/projected/a4f03f8d-fa97-450c-9117-fd601115c842-kube-api-access-bgf58\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.863017 4726 scope.go:117] "RemoveContainer" containerID="ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d" Nov 23 20:43:09 crc kubenswrapper[4726]: E1123 20:43:09.863597 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d\": container with ID starting with ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d not found: ID does not exist" containerID="ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.863638 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d"} err="failed to get container status \"ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d\": rpc error: code = NotFound desc = could not find container \"ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d\": container with ID starting with ec6f61d0ca6cba59f102394da0d484e28770fd59c14d3a5b6a363af67ba1234d not found: ID does not exist" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.863663 4726 scope.go:117] "RemoveContainer" containerID="b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd" Nov 23 20:43:09 crc kubenswrapper[4726]: E1123 20:43:09.864165 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd\": container with ID starting with b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd not found: ID does not exist" containerID="b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.864188 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd"} err="failed to get container status \"b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd\": rpc error: code = NotFound desc = could not find container \"b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd\": container with ID starting with b109a8a0f4c06ba2047fac5237aeb11d81b74127c4436b848b8c3215e10f69fd not found: ID does not exist" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.864203 4726 scope.go:117] "RemoveContainer" containerID="84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e" Nov 23 20:43:09 crc kubenswrapper[4726]: E1123 20:43:09.865741 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e\": container with ID starting with 84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e not found: ID does not exist" containerID="84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e" Nov 23 20:43:09 crc kubenswrapper[4726]: I1123 20:43:09.865764 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e"} err="failed to get container status \"84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e\": rpc error: code = NotFound desc = could not find container \"84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e\": container with ID starting with 84f87a666c0481c459865550c66058d39c4a6f46cd4871a8c9b56d152229c17e not found: ID does not exist" Nov 23 20:43:10 crc kubenswrapper[4726]: I1123 20:43:10.078287 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qvrg"] Nov 23 20:43:10 crc kubenswrapper[4726]: I1123 20:43:10.085740 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9qvrg"] Nov 23 20:43:10 crc kubenswrapper[4726]: I1123 20:43:10.606241 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" path="/var/lib/kubelet/pods/a4f03f8d-fa97-450c-9117-fd601115c842/volumes" Nov 23 20:43:13 crc kubenswrapper[4726]: I1123 20:43:13.104491 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:43:13 crc kubenswrapper[4726]: I1123 20:43:13.159648 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:43:13 crc kubenswrapper[4726]: I1123 20:43:13.888093 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kh9r6"] Nov 23 20:43:14 crc kubenswrapper[4726]: I1123 20:43:14.801579 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kh9r6" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="registry-server" containerID="cri-o://2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3" gracePeriod=2 Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.230486 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.351192 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-catalog-content\") pod \"f8492bea-d49a-4832-abf4-9964651479cc\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.351316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-utilities\") pod \"f8492bea-d49a-4832-abf4-9964651479cc\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.351523 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t4zv\" (UniqueName: \"kubernetes.io/projected/f8492bea-d49a-4832-abf4-9964651479cc-kube-api-access-8t4zv\") pod \"f8492bea-d49a-4832-abf4-9964651479cc\" (UID: \"f8492bea-d49a-4832-abf4-9964651479cc\") " Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.352006 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-utilities" (OuterVolumeSpecName: "utilities") pod "f8492bea-d49a-4832-abf4-9964651479cc" (UID: "f8492bea-d49a-4832-abf4-9964651479cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.352778 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.364648 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8492bea-d49a-4832-abf4-9964651479cc-kube-api-access-8t4zv" (OuterVolumeSpecName: "kube-api-access-8t4zv") pod "f8492bea-d49a-4832-abf4-9964651479cc" (UID: "f8492bea-d49a-4832-abf4-9964651479cc"). InnerVolumeSpecName "kube-api-access-8t4zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.445361 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8492bea-d49a-4832-abf4-9964651479cc" (UID: "f8492bea-d49a-4832-abf4-9964651479cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.454339 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t4zv\" (UniqueName: \"kubernetes.io/projected/f8492bea-d49a-4832-abf4-9964651479cc-kube-api-access-8t4zv\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.454380 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8492bea-d49a-4832-abf4-9964651479cc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.813097 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kh9r6" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.813956 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh9r6" event={"ID":"f8492bea-d49a-4832-abf4-9964651479cc","Type":"ContainerDied","Data":"2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3"} Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.814101 4726 scope.go:117] "RemoveContainer" containerID="2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.814038 4726 generic.go:334] "Generic (PLEG): container finished" podID="f8492bea-d49a-4832-abf4-9964651479cc" containerID="2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3" exitCode=0 Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.814327 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh9r6" event={"ID":"f8492bea-d49a-4832-abf4-9964651479cc","Type":"ContainerDied","Data":"5344c909f3992523e105160e752d38772c7221052a75a6d18c8cd9ec780635e3"} Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.831132 4726 scope.go:117] "RemoveContainer" containerID="1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.861119 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kh9r6"] Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.862369 4726 scope.go:117] "RemoveContainer" containerID="a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.869144 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kh9r6"] Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.892245 4726 scope.go:117] "RemoveContainer" containerID="2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3" Nov 23 20:43:15 crc kubenswrapper[4726]: E1123 20:43:15.892732 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3\": container with ID starting with 2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3 not found: ID does not exist" containerID="2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.892828 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3"} err="failed to get container status \"2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3\": rpc error: code = NotFound desc = could not find container \"2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3\": container with ID starting with 2af5b6342ab27022a5f131592e63aa25f55559eb1347406f163227a8f68687b3 not found: ID does not exist" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.893052 4726 scope.go:117] "RemoveContainer" containerID="1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48" Nov 23 20:43:15 crc kubenswrapper[4726]: E1123 20:43:15.893401 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48\": container with ID starting with 1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48 not found: ID does not exist" containerID="1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.893436 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48"} err="failed to get container status \"1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48\": rpc error: code = NotFound desc = could not find container \"1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48\": container with ID starting with 1e94a5fd5418d1deb0cf6fa54cd1196a0775d4d48436cddc37f61059256b0e48 not found: ID does not exist" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.893457 4726 scope.go:117] "RemoveContainer" containerID="a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17" Nov 23 20:43:15 crc kubenswrapper[4726]: E1123 20:43:15.893713 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17\": container with ID starting with a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17 not found: ID does not exist" containerID="a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17" Nov 23 20:43:15 crc kubenswrapper[4726]: I1123 20:43:15.893794 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17"} err="failed to get container status \"a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17\": rpc error: code = NotFound desc = could not find container \"a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17\": container with ID starting with a67c1cd506e2b1ac9a9e2aaeb830dffcf8469c8e3e86aab4e4a9e8a2ec5f3e17 not found: ID does not exist" Nov 23 20:43:16 crc kubenswrapper[4726]: I1123 20:43:16.601286 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8492bea-d49a-4832-abf4-9964651479cc" path="/var/lib/kubelet/pods/f8492bea-d49a-4832-abf4-9964651479cc/volumes" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.189445 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rg55t"] Nov 23 20:43:29 crc kubenswrapper[4726]: E1123 20:43:29.191888 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="extract-utilities" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.192045 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="extract-utilities" Nov 23 20:43:29 crc kubenswrapper[4726]: E1123 20:43:29.192159 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" containerName="extract-content" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.192262 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" containerName="extract-content" Nov 23 20:43:29 crc kubenswrapper[4726]: E1123 20:43:29.192378 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="extract-content" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.192498 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="extract-content" Nov 23 20:43:29 crc kubenswrapper[4726]: E1123 20:43:29.192619 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" containerName="registry-server" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.192726 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" containerName="registry-server" Nov 23 20:43:29 crc kubenswrapper[4726]: E1123 20:43:29.192852 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" containerName="extract-utilities" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.192995 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" containerName="extract-utilities" Nov 23 20:43:29 crc kubenswrapper[4726]: E1123 20:43:29.193126 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="registry-server" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.193236 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="registry-server" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.196806 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4f03f8d-fa97-450c-9117-fd601115c842" containerName="registry-server" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.197156 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8492bea-d49a-4832-abf4-9964651479cc" containerName="registry-server" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.199365 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.214382 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rg55t"] Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.325037 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-catalog-content\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.325096 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-utilities\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.325273 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lz59\" (UniqueName: \"kubernetes.io/projected/2ace9873-f572-4ab3-b80c-93aff47e6a8d-kube-api-access-8lz59\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.426679 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-catalog-content\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.426725 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-utilities\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.426783 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lz59\" (UniqueName: \"kubernetes.io/projected/2ace9873-f572-4ab3-b80c-93aff47e6a8d-kube-api-access-8lz59\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.427351 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-catalog-content\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.427379 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-utilities\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.461840 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lz59\" (UniqueName: \"kubernetes.io/projected/2ace9873-f572-4ab3-b80c-93aff47e6a8d-kube-api-access-8lz59\") pod \"community-operators-rg55t\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:29 crc kubenswrapper[4726]: I1123 20:43:29.521035 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.103740 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rg55t"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.735149 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.742346 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-srr4b"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.748599 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.754544 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.762063 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.767858 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.783961 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.784249 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.784334 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.801896 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnbw4"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.802169 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.802233 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-44l9d"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.810832 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xncpc"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.819624 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-srr4b"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.827032 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9vvmw"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.833594 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jnt59"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.844186 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pm78k"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.850961 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fhfmk"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.855096 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6fnz"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.860483 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-27vw5"] Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.964771 4726 generic.go:334] "Generic (PLEG): container finished" podID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerID="48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174" exitCode=0 Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.964852 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg55t" event={"ID":"2ace9873-f572-4ab3-b80c-93aff47e6a8d","Type":"ContainerDied","Data":"48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174"} Nov 23 20:43:30 crc kubenswrapper[4726]: I1123 20:43:30.964923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg55t" event={"ID":"2ace9873-f572-4ab3-b80c-93aff47e6a8d","Type":"ContainerStarted","Data":"619fae5d3b6ba1ec6f3e092955237a356fc18ba93256a4013ba9b2484bd3e4a5"} Nov 23 20:43:31 crc kubenswrapper[4726]: I1123 20:43:31.972989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg55t" event={"ID":"2ace9873-f572-4ab3-b80c-93aff47e6a8d","Type":"ContainerStarted","Data":"37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0"} Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.601367 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00328a0d-972a-4fcf-9889-58fbac5548b7" path="/var/lib/kubelet/pods/00328a0d-972a-4fcf-9889-58fbac5548b7/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.602373 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ff7e920-cb8b-453e-82cb-454e59325a43" path="/var/lib/kubelet/pods/0ff7e920-cb8b-453e-82cb-454e59325a43/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.603099 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ad4f10a-15ed-4107-9c4e-16975854883e" path="/var/lib/kubelet/pods/2ad4f10a-15ed-4107-9c4e-16975854883e/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.603750 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40ad6dc8-1508-4186-bbae-a5ff96044133" path="/var/lib/kubelet/pods/40ad6dc8-1508-4186-bbae-a5ff96044133/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.605103 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e36024a-9586-439f-b551-d5baba64775b" path="/var/lib/kubelet/pods/4e36024a-9586-439f-b551-d5baba64775b/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.605759 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c2cb05b-a398-4e68-b9e3-b660e47ce4d6" path="/var/lib/kubelet/pods/6c2cb05b-a398-4e68-b9e3-b660e47ce4d6/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.606420 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70977db3-0e6c-4718-bebe-1d14ad030b9d" path="/var/lib/kubelet/pods/70977db3-0e6c-4718-bebe-1d14ad030b9d/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.607632 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acbca385-261a-4dd2-98e3-602ee283e043" path="/var/lib/kubelet/pods/acbca385-261a-4dd2-98e3-602ee283e043/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.608302 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c23e6d0b-0b2d-43cd-96f1-556499de1202" path="/var/lib/kubelet/pods/c23e6d0b-0b2d-43cd-96f1-556499de1202/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.608977 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8139270-4ff9-440e-b998-aea1493900c4" path="/var/lib/kubelet/pods/e8139270-4ff9-440e-b998-aea1493900c4/volumes" Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.982156 4726 generic.go:334] "Generic (PLEG): container finished" podID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerID="37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0" exitCode=0 Nov 23 20:43:32 crc kubenswrapper[4726]: I1123 20:43:32.982227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg55t" event={"ID":"2ace9873-f572-4ab3-b80c-93aff47e6a8d","Type":"ContainerDied","Data":"37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0"} Nov 23 20:43:33 crc kubenswrapper[4726]: I1123 20:43:33.996465 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg55t" event={"ID":"2ace9873-f572-4ab3-b80c-93aff47e6a8d","Type":"ContainerStarted","Data":"81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4"} Nov 23 20:43:34 crc kubenswrapper[4726]: I1123 20:43:34.028989 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rg55t" podStartSLOduration=2.625666071 podStartE2EDuration="5.028970286s" podCreationTimestamp="2025-11-23 20:43:29 +0000 UTC" firstStartedPulling="2025-11-23 20:43:30.967214055 +0000 UTC m=+2119.116255041" lastFinishedPulling="2025-11-23 20:43:33.37051829 +0000 UTC m=+2121.519559256" observedRunningTime="2025-11-23 20:43:34.023733168 +0000 UTC m=+2122.172774144" watchObservedRunningTime="2025-11-23 20:43:34.028970286 +0000 UTC m=+2122.178011242" Nov 23 20:43:39 crc kubenswrapper[4726]: I1123 20:43:39.043018 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:43:39 crc kubenswrapper[4726]: I1123 20:43:39.043629 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:43:39 crc kubenswrapper[4726]: I1123 20:43:39.522174 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:39 crc kubenswrapper[4726]: I1123 20:43:39.523541 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:39 crc kubenswrapper[4726]: I1123 20:43:39.578441 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:40 crc kubenswrapper[4726]: I1123 20:43:40.093994 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:40 crc kubenswrapper[4726]: I1123 20:43:40.141991 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rg55t"] Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.057281 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rg55t" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerName="registry-server" containerID="cri-o://81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4" gracePeriod=2 Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.516511 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.583162 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-utilities\") pod \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.583257 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-catalog-content\") pod \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.583378 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lz59\" (UniqueName: \"kubernetes.io/projected/2ace9873-f572-4ab3-b80c-93aff47e6a8d-kube-api-access-8lz59\") pod \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\" (UID: \"2ace9873-f572-4ab3-b80c-93aff47e6a8d\") " Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.584410 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-utilities" (OuterVolumeSpecName: "utilities") pod "2ace9873-f572-4ab3-b80c-93aff47e6a8d" (UID: "2ace9873-f572-4ab3-b80c-93aff47e6a8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.607158 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ace9873-f572-4ab3-b80c-93aff47e6a8d-kube-api-access-8lz59" (OuterVolumeSpecName: "kube-api-access-8lz59") pod "2ace9873-f572-4ab3-b80c-93aff47e6a8d" (UID: "2ace9873-f572-4ab3-b80c-93aff47e6a8d"). InnerVolumeSpecName "kube-api-access-8lz59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.635206 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ace9873-f572-4ab3-b80c-93aff47e6a8d" (UID: "2ace9873-f572-4ab3-b80c-93aff47e6a8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.686624 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.686667 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ace9873-f572-4ab3-b80c-93aff47e6a8d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:42 crc kubenswrapper[4726]: I1123 20:43:42.686683 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lz59\" (UniqueName: \"kubernetes.io/projected/2ace9873-f572-4ab3-b80c-93aff47e6a8d-kube-api-access-8lz59\") on node \"crc\" DevicePath \"\"" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.072533 4726 generic.go:334] "Generic (PLEG): container finished" podID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerID="81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4" exitCode=0 Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.072604 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg55t" event={"ID":"2ace9873-f572-4ab3-b80c-93aff47e6a8d","Type":"ContainerDied","Data":"81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4"} Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.072638 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rg55t" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.072655 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rg55t" event={"ID":"2ace9873-f572-4ab3-b80c-93aff47e6a8d","Type":"ContainerDied","Data":"619fae5d3b6ba1ec6f3e092955237a356fc18ba93256a4013ba9b2484bd3e4a5"} Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.072685 4726 scope.go:117] "RemoveContainer" containerID="81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.098003 4726 scope.go:117] "RemoveContainer" containerID="37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.121050 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rg55t"] Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.130591 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rg55t"] Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.130860 4726 scope.go:117] "RemoveContainer" containerID="48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.172559 4726 scope.go:117] "RemoveContainer" containerID="81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4" Nov 23 20:43:43 crc kubenswrapper[4726]: E1123 20:43:43.172991 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4\": container with ID starting with 81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4 not found: ID does not exist" containerID="81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.173029 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4"} err="failed to get container status \"81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4\": rpc error: code = NotFound desc = could not find container \"81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4\": container with ID starting with 81d126c378b04b39fbd52c1527cc7be6754244422f60eb17b266fc4384f7a1b4 not found: ID does not exist" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.173053 4726 scope.go:117] "RemoveContainer" containerID="37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0" Nov 23 20:43:43 crc kubenswrapper[4726]: E1123 20:43:43.173332 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0\": container with ID starting with 37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0 not found: ID does not exist" containerID="37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.173380 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0"} err="failed to get container status \"37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0\": rpc error: code = NotFound desc = could not find container \"37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0\": container with ID starting with 37f72986794d4c7b8ee5345c4cdc77f3f04bf0387c30eeed53547a41edf730a0 not found: ID does not exist" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.173455 4726 scope.go:117] "RemoveContainer" containerID="48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174" Nov 23 20:43:43 crc kubenswrapper[4726]: E1123 20:43:43.173729 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174\": container with ID starting with 48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174 not found: ID does not exist" containerID="48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.173756 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174"} err="failed to get container status \"48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174\": rpc error: code = NotFound desc = could not find container \"48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174\": container with ID starting with 48fdedb89681075c91bacd454153ad4aaef09efccc70f1bf0ca7bfe9c0db1174 not found: ID does not exist" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.860644 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb"] Nov 23 20:43:43 crc kubenswrapper[4726]: E1123 20:43:43.861398 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerName="extract-utilities" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.861417 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerName="extract-utilities" Nov 23 20:43:43 crc kubenswrapper[4726]: E1123 20:43:43.861441 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerName="registry-server" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.861450 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerName="registry-server" Nov 23 20:43:43 crc kubenswrapper[4726]: E1123 20:43:43.861473 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerName="extract-content" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.861480 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerName="extract-content" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.861728 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" containerName="registry-server" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.862477 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.872500 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb"] Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.875826 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.875989 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.876069 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.877338 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:43:43 crc kubenswrapper[4726]: I1123 20:43:43.877506 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.011076 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.011129 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.011263 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlclz\" (UniqueName: \"kubernetes.io/projected/dc981c46-b927-4734-a0d1-061826f92146-kube-api-access-hlclz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.011326 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.011380 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.112746 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlclz\" (UniqueName: \"kubernetes.io/projected/dc981c46-b927-4734-a0d1-061826f92146-kube-api-access-hlclz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.112806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.112834 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.112934 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.113593 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.118342 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.119472 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.119151 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.120838 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.137548 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlclz\" (UniqueName: \"kubernetes.io/projected/dc981c46-b927-4734-a0d1-061826f92146-kube-api-access-hlclz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.184577 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.600330 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ace9873-f572-4ab3-b80c-93aff47e6a8d" path="/var/lib/kubelet/pods/2ace9873-f572-4ab3-b80c-93aff47e6a8d/volumes" Nov 23 20:43:44 crc kubenswrapper[4726]: I1123 20:43:44.744684 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb"] Nov 23 20:43:45 crc kubenswrapper[4726]: I1123 20:43:45.091914 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" event={"ID":"dc981c46-b927-4734-a0d1-061826f92146","Type":"ContainerStarted","Data":"7b020de70d34e43665873b6150e24125dd361644ff5870f3523c6bded88c6d16"} Nov 23 20:43:46 crc kubenswrapper[4726]: I1123 20:43:46.099571 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" event={"ID":"dc981c46-b927-4734-a0d1-061826f92146","Type":"ContainerStarted","Data":"925bc92aed6d349830ff74b0a075a15d64464f799eb1714a54203cd064200e89"} Nov 23 20:43:46 crc kubenswrapper[4726]: I1123 20:43:46.121557 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" podStartSLOduration=2.467143735 podStartE2EDuration="3.121529098s" podCreationTimestamp="2025-11-23 20:43:43 +0000 UTC" firstStartedPulling="2025-11-23 20:43:44.742926867 +0000 UTC m=+2132.891967823" lastFinishedPulling="2025-11-23 20:43:45.39731223 +0000 UTC m=+2133.546353186" observedRunningTime="2025-11-23 20:43:46.113679056 +0000 UTC m=+2134.262720012" watchObservedRunningTime="2025-11-23 20:43:46.121529098 +0000 UTC m=+2134.270570094" Nov 23 20:44:02 crc kubenswrapper[4726]: I1123 20:44:02.288322 4726 generic.go:334] "Generic (PLEG): container finished" podID="dc981c46-b927-4734-a0d1-061826f92146" containerID="925bc92aed6d349830ff74b0a075a15d64464f799eb1714a54203cd064200e89" exitCode=0 Nov 23 20:44:02 crc kubenswrapper[4726]: I1123 20:44:02.289977 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" event={"ID":"dc981c46-b927-4734-a0d1-061826f92146","Type":"ContainerDied","Data":"925bc92aed6d349830ff74b0a075a15d64464f799eb1714a54203cd064200e89"} Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.816460 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.836232 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-repo-setup-combined-ca-bundle\") pod \"dc981c46-b927-4734-a0d1-061826f92146\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.836303 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ssh-key\") pod \"dc981c46-b927-4734-a0d1-061826f92146\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.836373 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-inventory\") pod \"dc981c46-b927-4734-a0d1-061826f92146\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.836410 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlclz\" (UniqueName: \"kubernetes.io/projected/dc981c46-b927-4734-a0d1-061826f92146-kube-api-access-hlclz\") pod \"dc981c46-b927-4734-a0d1-061826f92146\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.836469 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ceph\") pod \"dc981c46-b927-4734-a0d1-061826f92146\" (UID: \"dc981c46-b927-4734-a0d1-061826f92146\") " Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.843062 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc981c46-b927-4734-a0d1-061826f92146-kube-api-access-hlclz" (OuterVolumeSpecName: "kube-api-access-hlclz") pod "dc981c46-b927-4734-a0d1-061826f92146" (UID: "dc981c46-b927-4734-a0d1-061826f92146"). InnerVolumeSpecName "kube-api-access-hlclz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.855491 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "dc981c46-b927-4734-a0d1-061826f92146" (UID: "dc981c46-b927-4734-a0d1-061826f92146"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.858164 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ceph" (OuterVolumeSpecName: "ceph") pod "dc981c46-b927-4734-a0d1-061826f92146" (UID: "dc981c46-b927-4734-a0d1-061826f92146"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.871020 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dc981c46-b927-4734-a0d1-061826f92146" (UID: "dc981c46-b927-4734-a0d1-061826f92146"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.876074 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-inventory" (OuterVolumeSpecName: "inventory") pod "dc981c46-b927-4734-a0d1-061826f92146" (UID: "dc981c46-b927-4734-a0d1-061826f92146"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.937830 4726 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.937859 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.937882 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.937892 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlclz\" (UniqueName: \"kubernetes.io/projected/dc981c46-b927-4734-a0d1-061826f92146-kube-api-access-hlclz\") on node \"crc\" DevicePath \"\"" Nov 23 20:44:03 crc kubenswrapper[4726]: I1123 20:44:03.937903 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dc981c46-b927-4734-a0d1-061826f92146-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.317156 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" event={"ID":"dc981c46-b927-4734-a0d1-061826f92146","Type":"ContainerDied","Data":"7b020de70d34e43665873b6150e24125dd361644ff5870f3523c6bded88c6d16"} Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.318151 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b020de70d34e43665873b6150e24125dd361644ff5870f3523c6bded88c6d16" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.317241 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.426632 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7"] Nov 23 20:44:04 crc kubenswrapper[4726]: E1123 20:44:04.427079 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc981c46-b927-4734-a0d1-061826f92146" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.427104 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc981c46-b927-4734-a0d1-061826f92146" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.427317 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc981c46-b927-4734-a0d1-061826f92146" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.428009 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.431517 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.431849 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.432047 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.432193 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.438515 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.441112 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7"] Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.447745 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpdsb\" (UniqueName: \"kubernetes.io/projected/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-kube-api-access-fpdsb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.447799 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.447846 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.447885 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.447953 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.551089 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.551175 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.551320 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.551419 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpdsb\" (UniqueName: \"kubernetes.io/projected/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-kube-api-access-fpdsb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.551493 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.557762 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.558236 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.560580 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.562460 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.575901 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpdsb\" (UniqueName: \"kubernetes.io/projected/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-kube-api-access-fpdsb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:04 crc kubenswrapper[4726]: I1123 20:44:04.749019 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:44:05 crc kubenswrapper[4726]: I1123 20:44:05.310325 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7"] Nov 23 20:44:05 crc kubenswrapper[4726]: I1123 20:44:05.331723 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" event={"ID":"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3","Type":"ContainerStarted","Data":"57ea379711bf6a99e24bd9edc91de599883a93fb827e913551730dad15d30dc1"} Nov 23 20:44:06 crc kubenswrapper[4726]: I1123 20:44:06.339526 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" event={"ID":"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3","Type":"ContainerStarted","Data":"18669ea24aa7e9ce8685302855b0236afb80160c04e6bd887f77a7433a803937"} Nov 23 20:44:09 crc kubenswrapper[4726]: I1123 20:44:09.043353 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:44:09 crc kubenswrapper[4726]: I1123 20:44:09.043751 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.282769 4726 scope.go:117] "RemoveContainer" containerID="8cd5808554f84ac6456cde393aed0c648e4a4e329828651bc572e8dcc884b9a6" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.338078 4726 scope.go:117] "RemoveContainer" containerID="5b131b0e71d29f3c72f90ecf13a46837967d4af91e9c038293dbfc781f195501" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.372987 4726 scope.go:117] "RemoveContainer" containerID="2553283fdb55c90da0ab0b30e350932a587049af4077cf26d70079a6397a7ad2" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.421705 4726 scope.go:117] "RemoveContainer" containerID="655e80a4c0e6045c113f4cf5bca2964a39a0cf2a5fdd8f7fcfbe6fa26abb58c1" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.450682 4726 scope.go:117] "RemoveContainer" containerID="740fcead03fe7884ee89023a0f11ad53bdc4e850757cd8a8a92051626d94f729" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.559209 4726 scope.go:117] "RemoveContainer" containerID="12999a5709031f4664e1f46234d94e7e243fb971f53e77b58ba815d48d285a98" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.586380 4726 scope.go:117] "RemoveContainer" containerID="ae51844d8d85283811611f7344fbf3bac3c0353ff46cc259a81987ca4964d54f" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.624402 4726 scope.go:117] "RemoveContainer" containerID="6279b72420a8ef5e3fd08686627f37edef6851aad393e3a83b76912dba0ce494" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.696798 4726 scope.go:117] "RemoveContainer" containerID="c7c2efa2d1aa7d18e62b6de42d411a1e5e1c671d1271cb47d13b2aba9c043cb7" Nov 23 20:44:19 crc kubenswrapper[4726]: I1123 20:44:19.726805 4726 scope.go:117] "RemoveContainer" containerID="4f9241e8934dc5253b19735fd14ccfc3b88975a33e356abf2adf5d0d66a18e64" Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.042665 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.043373 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.043423 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.044031 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.044097 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" gracePeriod=600 Nov 23 20:44:39 crc kubenswrapper[4726]: E1123 20:44:39.175311 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.643852 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" exitCode=0 Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.644206 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e"} Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.644481 4726 scope.go:117] "RemoveContainer" containerID="b4944e760d4a10e30b574916d9b12b8294572948eb1263ca192e2958218c0945" Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.645470 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:44:39 crc kubenswrapper[4726]: E1123 20:44:39.645931 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:44:39 crc kubenswrapper[4726]: I1123 20:44:39.668554 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" podStartSLOduration=35.255834706 podStartE2EDuration="35.668528437s" podCreationTimestamp="2025-11-23 20:44:04 +0000 UTC" firstStartedPulling="2025-11-23 20:44:05.32010774 +0000 UTC m=+2153.469148716" lastFinishedPulling="2025-11-23 20:44:05.732801491 +0000 UTC m=+2153.881842447" observedRunningTime="2025-11-23 20:44:06.359597661 +0000 UTC m=+2154.508638627" watchObservedRunningTime="2025-11-23 20:44:39.668528437 +0000 UTC m=+2187.817569403" Nov 23 20:44:53 crc kubenswrapper[4726]: I1123 20:44:53.590501 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:44:53 crc kubenswrapper[4726]: E1123 20:44:53.592504 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.169729 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw"] Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.175015 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.181857 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.182223 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.190011 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw"] Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.199271 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8r84\" (UniqueName: \"kubernetes.io/projected/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-kube-api-access-p8r84\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.199393 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-config-volume\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.199478 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-secret-volume\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.301838 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-config-volume\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.301937 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-secret-volume\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.302055 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8r84\" (UniqueName: \"kubernetes.io/projected/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-kube-api-access-p8r84\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.303205 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-config-volume\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.309667 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-secret-volume\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.321793 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8r84\" (UniqueName: \"kubernetes.io/projected/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-kube-api-access-p8r84\") pod \"collect-profiles-29398845-cjtfw\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:00 crc kubenswrapper[4726]: I1123 20:45:00.522154 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:01 crc kubenswrapper[4726]: I1123 20:45:01.008479 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw"] Nov 23 20:45:01 crc kubenswrapper[4726]: W1123 20:45:01.016463 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10ad4a6d_d9fa_47b9_8b5d_1c5df7269a31.slice/crio-0a8c439ef7311f7654a0d66d0d7b6ce3ebd2507db93a47053a7b6ec7c6407076 WatchSource:0}: Error finding container 0a8c439ef7311f7654a0d66d0d7b6ce3ebd2507db93a47053a7b6ec7c6407076: Status 404 returned error can't find the container with id 0a8c439ef7311f7654a0d66d0d7b6ce3ebd2507db93a47053a7b6ec7c6407076 Nov 23 20:45:01 crc kubenswrapper[4726]: I1123 20:45:01.883548 4726 generic.go:334] "Generic (PLEG): container finished" podID="10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31" containerID="2f21f64c85620c4c9f4799f6a51fc2c54515e2ff652a4f3053b1a57694e663bd" exitCode=0 Nov 23 20:45:01 crc kubenswrapper[4726]: I1123 20:45:01.883619 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" event={"ID":"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31","Type":"ContainerDied","Data":"2f21f64c85620c4c9f4799f6a51fc2c54515e2ff652a4f3053b1a57694e663bd"} Nov 23 20:45:01 crc kubenswrapper[4726]: I1123 20:45:01.884041 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" event={"ID":"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31","Type":"ContainerStarted","Data":"0a8c439ef7311f7654a0d66d0d7b6ce3ebd2507db93a47053a7b6ec7c6407076"} Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.295033 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.473060 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8r84\" (UniqueName: \"kubernetes.io/projected/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-kube-api-access-p8r84\") pod \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.473269 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-secret-volume\") pod \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.473380 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-config-volume\") pod \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\" (UID: \"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31\") " Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.474207 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-config-volume" (OuterVolumeSpecName: "config-volume") pod "10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31" (UID: "10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.474513 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.482185 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-kube-api-access-p8r84" (OuterVolumeSpecName: "kube-api-access-p8r84") pod "10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31" (UID: "10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31"). InnerVolumeSpecName "kube-api-access-p8r84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.482978 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31" (UID: "10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.576477 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8r84\" (UniqueName: \"kubernetes.io/projected/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-kube-api-access-p8r84\") on node \"crc\" DevicePath \"\"" Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.576515 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.905860 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" event={"ID":"10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31","Type":"ContainerDied","Data":"0a8c439ef7311f7654a0d66d0d7b6ce3ebd2507db93a47053a7b6ec7c6407076"} Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.905955 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a8c439ef7311f7654a0d66d0d7b6ce3ebd2507db93a47053a7b6ec7c6407076" Nov 23 20:45:03 crc kubenswrapper[4726]: I1123 20:45:03.905913 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw" Nov 23 20:45:04 crc kubenswrapper[4726]: I1123 20:45:04.383406 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h"] Nov 23 20:45:04 crc kubenswrapper[4726]: I1123 20:45:04.393560 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398800-x676h"] Nov 23 20:45:04 crc kubenswrapper[4726]: I1123 20:45:04.600891 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c969e63-8584-4613-85be-1d11f3014d13" path="/var/lib/kubelet/pods/8c969e63-8584-4613-85be-1d11f3014d13/volumes" Nov 23 20:45:05 crc kubenswrapper[4726]: I1123 20:45:05.589254 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:45:05 crc kubenswrapper[4726]: E1123 20:45:05.589583 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:45:18 crc kubenswrapper[4726]: I1123 20:45:18.589296 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:45:18 crc kubenswrapper[4726]: E1123 20:45:18.590299 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:45:19 crc kubenswrapper[4726]: I1123 20:45:19.924923 4726 scope.go:117] "RemoveContainer" containerID="a1936471dbabc26a0a6359a5ef2e5d10d2bb1338d573d418a8e49cb44f3567e5" Nov 23 20:45:32 crc kubenswrapper[4726]: I1123 20:45:32.595518 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:45:32 crc kubenswrapper[4726]: E1123 20:45:32.597349 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.283105 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k75q4"] Nov 23 20:45:46 crc kubenswrapper[4726]: E1123 20:45:46.284042 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31" containerName="collect-profiles" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.284054 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31" containerName="collect-profiles" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.284219 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31" containerName="collect-profiles" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.285514 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.297080 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k75q4"] Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.358861 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdp9z\" (UniqueName: \"kubernetes.io/projected/67908dff-2057-486f-8e10-caf29fd07a85-kube-api-access-xdp9z\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.358922 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-catalog-content\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.358995 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-utilities\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.460609 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdp9z\" (UniqueName: \"kubernetes.io/projected/67908dff-2057-486f-8e10-caf29fd07a85-kube-api-access-xdp9z\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.460993 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-catalog-content\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.461086 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-utilities\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.461592 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-utilities\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.462167 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-catalog-content\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.482924 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdp9z\" (UniqueName: \"kubernetes.io/projected/67908dff-2057-486f-8e10-caf29fd07a85-kube-api-access-xdp9z\") pod \"certified-operators-k75q4\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.589838 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:45:46 crc kubenswrapper[4726]: E1123 20:45:46.590154 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:45:46 crc kubenswrapper[4726]: I1123 20:45:46.638335 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:47 crc kubenswrapper[4726]: I1123 20:45:47.166335 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k75q4"] Nov 23 20:45:47 crc kubenswrapper[4726]: I1123 20:45:47.343677 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k75q4" event={"ID":"67908dff-2057-486f-8e10-caf29fd07a85","Type":"ContainerStarted","Data":"af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076"} Nov 23 20:45:47 crc kubenswrapper[4726]: I1123 20:45:47.343896 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k75q4" event={"ID":"67908dff-2057-486f-8e10-caf29fd07a85","Type":"ContainerStarted","Data":"056054ea5669458f378f4d602f550015e7a4b1caa540ffd1ab8845aae6a49af5"} Nov 23 20:45:48 crc kubenswrapper[4726]: I1123 20:45:48.351353 4726 generic.go:334] "Generic (PLEG): container finished" podID="67908dff-2057-486f-8e10-caf29fd07a85" containerID="af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076" exitCode=0 Nov 23 20:45:48 crc kubenswrapper[4726]: I1123 20:45:48.351660 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k75q4" event={"ID":"67908dff-2057-486f-8e10-caf29fd07a85","Type":"ContainerDied","Data":"af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076"} Nov 23 20:45:48 crc kubenswrapper[4726]: I1123 20:45:48.351689 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k75q4" event={"ID":"67908dff-2057-486f-8e10-caf29fd07a85","Type":"ContainerStarted","Data":"a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990"} Nov 23 20:45:50 crc kubenswrapper[4726]: I1123 20:45:50.368640 4726 generic.go:334] "Generic (PLEG): container finished" podID="67908dff-2057-486f-8e10-caf29fd07a85" containerID="a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990" exitCode=0 Nov 23 20:45:50 crc kubenswrapper[4726]: I1123 20:45:50.368705 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k75q4" event={"ID":"67908dff-2057-486f-8e10-caf29fd07a85","Type":"ContainerDied","Data":"a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990"} Nov 23 20:45:51 crc kubenswrapper[4726]: I1123 20:45:51.379860 4726 generic.go:334] "Generic (PLEG): container finished" podID="e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" containerID="18669ea24aa7e9ce8685302855b0236afb80160c04e6bd887f77a7433a803937" exitCode=0 Nov 23 20:45:51 crc kubenswrapper[4726]: I1123 20:45:51.379974 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" event={"ID":"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3","Type":"ContainerDied","Data":"18669ea24aa7e9ce8685302855b0236afb80160c04e6bd887f77a7433a803937"} Nov 23 20:45:51 crc kubenswrapper[4726]: I1123 20:45:51.382008 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k75q4" event={"ID":"67908dff-2057-486f-8e10-caf29fd07a85","Type":"ContainerStarted","Data":"e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520"} Nov 23 20:45:51 crc kubenswrapper[4726]: I1123 20:45:51.417632 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k75q4" podStartSLOduration=1.957646157 podStartE2EDuration="5.417616431s" podCreationTimestamp="2025-11-23 20:45:46 +0000 UTC" firstStartedPulling="2025-11-23 20:45:47.345214979 +0000 UTC m=+2255.494255925" lastFinishedPulling="2025-11-23 20:45:50.805185203 +0000 UTC m=+2258.954226199" observedRunningTime="2025-11-23 20:45:51.413439633 +0000 UTC m=+2259.562480589" watchObservedRunningTime="2025-11-23 20:45:51.417616431 +0000 UTC m=+2259.566657387" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.810755 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.870317 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpdsb\" (UniqueName: \"kubernetes.io/projected/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-kube-api-access-fpdsb\") pod \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.870378 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-inventory\") pod \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.870421 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ssh-key\") pod \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.870452 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ceph\") pod \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.870511 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-bootstrap-combined-ca-bundle\") pod \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\" (UID: \"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3\") " Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.876050 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" (UID: "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.876521 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ceph" (OuterVolumeSpecName: "ceph") pod "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" (UID: "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.876594 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-kube-api-access-fpdsb" (OuterVolumeSpecName: "kube-api-access-fpdsb") pod "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" (UID: "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3"). InnerVolumeSpecName "kube-api-access-fpdsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.897977 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" (UID: "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.910421 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-inventory" (OuterVolumeSpecName: "inventory") pod "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" (UID: "e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.972187 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.972384 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.972469 4726 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.972534 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpdsb\" (UniqueName: \"kubernetes.io/projected/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-kube-api-access-fpdsb\") on node \"crc\" DevicePath \"\"" Nov 23 20:45:52 crc kubenswrapper[4726]: I1123 20:45:52.972588 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.397092 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" event={"ID":"e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3","Type":"ContainerDied","Data":"57ea379711bf6a99e24bd9edc91de599883a93fb827e913551730dad15d30dc1"} Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.397142 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57ea379711bf6a99e24bd9edc91de599883a93fb827e913551730dad15d30dc1" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.397156 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.506665 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k"] Nov 23 20:45:53 crc kubenswrapper[4726]: E1123 20:45:53.507101 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.507124 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.507338 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.508073 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.511398 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.511712 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.513969 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.514486 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.514852 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.519375 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k"] Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.581787 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.581884 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.581935 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.581978 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44cwg\" (UniqueName: \"kubernetes.io/projected/74d7e759-c823-49e1-9480-87d93552cd4c-kube-api-access-44cwg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.684019 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.684080 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.684113 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.684163 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44cwg\" (UniqueName: \"kubernetes.io/projected/74d7e759-c823-49e1-9480-87d93552cd4c-kube-api-access-44cwg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.688385 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.688777 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.689327 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.698044 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44cwg\" (UniqueName: \"kubernetes.io/projected/74d7e759-c823-49e1-9480-87d93552cd4c-kube-api-access-44cwg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rd42k\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:53 crc kubenswrapper[4726]: I1123 20:45:53.841238 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:45:54 crc kubenswrapper[4726]: I1123 20:45:54.351298 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k"] Nov 23 20:45:54 crc kubenswrapper[4726]: I1123 20:45:54.406160 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" event={"ID":"74d7e759-c823-49e1-9480-87d93552cd4c","Type":"ContainerStarted","Data":"7148b99ac4300d261d12dd448725df47de3dd168eca23eab7c1296113d7a3007"} Nov 23 20:45:55 crc kubenswrapper[4726]: I1123 20:45:55.436389 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" event={"ID":"74d7e759-c823-49e1-9480-87d93552cd4c","Type":"ContainerStarted","Data":"380b50f546b87e86d86724ce2ba6f2810832f30c577f19c58bc2f2cc067c0cd7"} Nov 23 20:45:55 crc kubenswrapper[4726]: I1123 20:45:55.466423 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" podStartSLOduration=2.001604846 podStartE2EDuration="2.466405513s" podCreationTimestamp="2025-11-23 20:45:53 +0000 UTC" firstStartedPulling="2025-11-23 20:45:54.358631411 +0000 UTC m=+2262.507672397" lastFinishedPulling="2025-11-23 20:45:54.823432108 +0000 UTC m=+2262.972473064" observedRunningTime="2025-11-23 20:45:55.459536548 +0000 UTC m=+2263.608577504" watchObservedRunningTime="2025-11-23 20:45:55.466405513 +0000 UTC m=+2263.615446469" Nov 23 20:45:56 crc kubenswrapper[4726]: I1123 20:45:56.638965 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:56 crc kubenswrapper[4726]: I1123 20:45:56.639008 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:56 crc kubenswrapper[4726]: I1123 20:45:56.689891 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:57 crc kubenswrapper[4726]: I1123 20:45:57.505843 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:45:57 crc kubenswrapper[4726]: I1123 20:45:57.566244 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k75q4"] Nov 23 20:45:59 crc kubenswrapper[4726]: I1123 20:45:59.475504 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k75q4" podUID="67908dff-2057-486f-8e10-caf29fd07a85" containerName="registry-server" containerID="cri-o://e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520" gracePeriod=2 Nov 23 20:45:59 crc kubenswrapper[4726]: I1123 20:45:59.947123 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.140711 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-utilities\") pod \"67908dff-2057-486f-8e10-caf29fd07a85\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.140780 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-catalog-content\") pod \"67908dff-2057-486f-8e10-caf29fd07a85\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.140918 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdp9z\" (UniqueName: \"kubernetes.io/projected/67908dff-2057-486f-8e10-caf29fd07a85-kube-api-access-xdp9z\") pod \"67908dff-2057-486f-8e10-caf29fd07a85\" (UID: \"67908dff-2057-486f-8e10-caf29fd07a85\") " Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.143233 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-utilities" (OuterVolumeSpecName: "utilities") pod "67908dff-2057-486f-8e10-caf29fd07a85" (UID: "67908dff-2057-486f-8e10-caf29fd07a85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.150102 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67908dff-2057-486f-8e10-caf29fd07a85-kube-api-access-xdp9z" (OuterVolumeSpecName: "kube-api-access-xdp9z") pod "67908dff-2057-486f-8e10-caf29fd07a85" (UID: "67908dff-2057-486f-8e10-caf29fd07a85"). InnerVolumeSpecName "kube-api-access-xdp9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.225714 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67908dff-2057-486f-8e10-caf29fd07a85" (UID: "67908dff-2057-486f-8e10-caf29fd07a85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.243392 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.243419 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67908dff-2057-486f-8e10-caf29fd07a85-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.243430 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdp9z\" (UniqueName: \"kubernetes.io/projected/67908dff-2057-486f-8e10-caf29fd07a85-kube-api-access-xdp9z\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.485210 4726 generic.go:334] "Generic (PLEG): container finished" podID="67908dff-2057-486f-8e10-caf29fd07a85" containerID="e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520" exitCode=0 Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.485260 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k75q4" event={"ID":"67908dff-2057-486f-8e10-caf29fd07a85","Type":"ContainerDied","Data":"e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520"} Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.485271 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k75q4" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.485288 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k75q4" event={"ID":"67908dff-2057-486f-8e10-caf29fd07a85","Type":"ContainerDied","Data":"056054ea5669458f378f4d602f550015e7a4b1caa540ffd1ab8845aae6a49af5"} Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.485309 4726 scope.go:117] "RemoveContainer" containerID="e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.505630 4726 scope.go:117] "RemoveContainer" containerID="a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.544621 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k75q4"] Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.546151 4726 scope.go:117] "RemoveContainer" containerID="af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.551819 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k75q4"] Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.590153 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:46:00 crc kubenswrapper[4726]: E1123 20:46:00.590440 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.591713 4726 scope.go:117] "RemoveContainer" containerID="e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520" Nov 23 20:46:00 crc kubenswrapper[4726]: E1123 20:46:00.592761 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520\": container with ID starting with e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520 not found: ID does not exist" containerID="e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.592793 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520"} err="failed to get container status \"e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520\": rpc error: code = NotFound desc = could not find container \"e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520\": container with ID starting with e2e42e0ff17d9768bf07edeedac47b06bd7c41deb33c3dbea5194ec446ea0520 not found: ID does not exist" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.592816 4726 scope.go:117] "RemoveContainer" containerID="a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990" Nov 23 20:46:00 crc kubenswrapper[4726]: E1123 20:46:00.593259 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990\": container with ID starting with a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990 not found: ID does not exist" containerID="a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.593281 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990"} err="failed to get container status \"a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990\": rpc error: code = NotFound desc = could not find container \"a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990\": container with ID starting with a94423993d1af8b774a50a6b35f618bbe07ade4b21ea7b8a9be4788c43ea2990 not found: ID does not exist" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.593295 4726 scope.go:117] "RemoveContainer" containerID="af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076" Nov 23 20:46:00 crc kubenswrapper[4726]: E1123 20:46:00.593510 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076\": container with ID starting with af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076 not found: ID does not exist" containerID="af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.593527 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076"} err="failed to get container status \"af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076\": rpc error: code = NotFound desc = could not find container \"af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076\": container with ID starting with af6aa4dd3eb5936636f050d8277ec2d80dcdbb72e13ba9b7186c6d2a8a074076 not found: ID does not exist" Nov 23 20:46:00 crc kubenswrapper[4726]: I1123 20:46:00.601896 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67908dff-2057-486f-8e10-caf29fd07a85" path="/var/lib/kubelet/pods/67908dff-2057-486f-8e10-caf29fd07a85/volumes" Nov 23 20:46:14 crc kubenswrapper[4726]: I1123 20:46:14.589737 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:46:14 crc kubenswrapper[4726]: E1123 20:46:14.591372 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:46:23 crc kubenswrapper[4726]: I1123 20:46:23.725267 4726 generic.go:334] "Generic (PLEG): container finished" podID="74d7e759-c823-49e1-9480-87d93552cd4c" containerID="380b50f546b87e86d86724ce2ba6f2810832f30c577f19c58bc2f2cc067c0cd7" exitCode=0 Nov 23 20:46:23 crc kubenswrapper[4726]: I1123 20:46:23.725997 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" event={"ID":"74d7e759-c823-49e1-9480-87d93552cd4c","Type":"ContainerDied","Data":"380b50f546b87e86d86724ce2ba6f2810832f30c577f19c58bc2f2cc067c0cd7"} Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.173829 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.185631 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-inventory\") pod \"74d7e759-c823-49e1-9480-87d93552cd4c\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.185764 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44cwg\" (UniqueName: \"kubernetes.io/projected/74d7e759-c823-49e1-9480-87d93552cd4c-kube-api-access-44cwg\") pod \"74d7e759-c823-49e1-9480-87d93552cd4c\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.185807 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ceph\") pod \"74d7e759-c823-49e1-9480-87d93552cd4c\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.185851 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ssh-key\") pod \"74d7e759-c823-49e1-9480-87d93552cd4c\" (UID: \"74d7e759-c823-49e1-9480-87d93552cd4c\") " Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.200090 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ceph" (OuterVolumeSpecName: "ceph") pod "74d7e759-c823-49e1-9480-87d93552cd4c" (UID: "74d7e759-c823-49e1-9480-87d93552cd4c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.214471 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74d7e759-c823-49e1-9480-87d93552cd4c-kube-api-access-44cwg" (OuterVolumeSpecName: "kube-api-access-44cwg") pod "74d7e759-c823-49e1-9480-87d93552cd4c" (UID: "74d7e759-c823-49e1-9480-87d93552cd4c"). InnerVolumeSpecName "kube-api-access-44cwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.227069 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74d7e759-c823-49e1-9480-87d93552cd4c" (UID: "74d7e759-c823-49e1-9480-87d93552cd4c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.227187 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-inventory" (OuterVolumeSpecName: "inventory") pod "74d7e759-c823-49e1-9480-87d93552cd4c" (UID: "74d7e759-c823-49e1-9480-87d93552cd4c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.287755 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.287788 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44cwg\" (UniqueName: \"kubernetes.io/projected/74d7e759-c823-49e1-9480-87d93552cd4c-kube-api-access-44cwg\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.287800 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.287810 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74d7e759-c823-49e1-9480-87d93552cd4c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.750062 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" event={"ID":"74d7e759-c823-49e1-9480-87d93552cd4c","Type":"ContainerDied","Data":"7148b99ac4300d261d12dd448725df47de3dd168eca23eab7c1296113d7a3007"} Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.750125 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7148b99ac4300d261d12dd448725df47de3dd168eca23eab7c1296113d7a3007" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.750207 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rd42k" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.860612 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn"] Nov 23 20:46:25 crc kubenswrapper[4726]: E1123 20:46:25.860958 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67908dff-2057-486f-8e10-caf29fd07a85" containerName="registry-server" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.860973 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67908dff-2057-486f-8e10-caf29fd07a85" containerName="registry-server" Nov 23 20:46:25 crc kubenswrapper[4726]: E1123 20:46:25.860992 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67908dff-2057-486f-8e10-caf29fd07a85" containerName="extract-utilities" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.860999 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67908dff-2057-486f-8e10-caf29fd07a85" containerName="extract-utilities" Nov 23 20:46:25 crc kubenswrapper[4726]: E1123 20:46:25.861012 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67908dff-2057-486f-8e10-caf29fd07a85" containerName="extract-content" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.861018 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67908dff-2057-486f-8e10-caf29fd07a85" containerName="extract-content" Nov 23 20:46:25 crc kubenswrapper[4726]: E1123 20:46:25.861044 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74d7e759-c823-49e1-9480-87d93552cd4c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.861051 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="74d7e759-c823-49e1-9480-87d93552cd4c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.861217 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="67908dff-2057-486f-8e10-caf29fd07a85" containerName="registry-server" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.861229 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="74d7e759-c823-49e1-9480-87d93552cd4c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.861764 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.864137 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.864432 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.864585 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.865407 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.870483 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:46:25 crc kubenswrapper[4726]: I1123 20:46:25.883927 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn"] Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.005304 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.005377 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.005399 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.005616 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr6jv\" (UniqueName: \"kubernetes.io/projected/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-kube-api-access-mr6jv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.107422 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.107484 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.107502 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.107552 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr6jv\" (UniqueName: \"kubernetes.io/projected/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-kube-api-access-mr6jv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.112181 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.113963 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.115687 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.136293 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr6jv\" (UniqueName: \"kubernetes.io/projected/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-kube-api-access-mr6jv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.182260 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.737284 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn"] Nov 23 20:46:26 crc kubenswrapper[4726]: I1123 20:46:26.762301 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" event={"ID":"da03c2c2-f43a-4f51-9c87-8e2f6b0af367","Type":"ContainerStarted","Data":"4c1586d86646b190d71b4dace567af707674ce86215b6dd8cf184158785b5f73"} Nov 23 20:46:27 crc kubenswrapper[4726]: I1123 20:46:27.772251 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" event={"ID":"da03c2c2-f43a-4f51-9c87-8e2f6b0af367","Type":"ContainerStarted","Data":"8e9ebdba2b7ba70e002116caf37288c6ad9809835e0e3e9cd9c81dc7b7cf0d7f"} Nov 23 20:46:27 crc kubenswrapper[4726]: I1123 20:46:27.801913 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" podStartSLOduration=2.364330985 podStartE2EDuration="2.801893028s" podCreationTimestamp="2025-11-23 20:46:25 +0000 UTC" firstStartedPulling="2025-11-23 20:46:26.743355525 +0000 UTC m=+2294.892396481" lastFinishedPulling="2025-11-23 20:46:27.180917548 +0000 UTC m=+2295.329958524" observedRunningTime="2025-11-23 20:46:27.79524318 +0000 UTC m=+2295.944284126" watchObservedRunningTime="2025-11-23 20:46:27.801893028 +0000 UTC m=+2295.950933994" Nov 23 20:46:29 crc kubenswrapper[4726]: I1123 20:46:29.589073 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:46:29 crc kubenswrapper[4726]: E1123 20:46:29.589723 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:46:32 crc kubenswrapper[4726]: I1123 20:46:32.817346 4726 generic.go:334] "Generic (PLEG): container finished" podID="da03c2c2-f43a-4f51-9c87-8e2f6b0af367" containerID="8e9ebdba2b7ba70e002116caf37288c6ad9809835e0e3e9cd9c81dc7b7cf0d7f" exitCode=0 Nov 23 20:46:32 crc kubenswrapper[4726]: I1123 20:46:32.817451 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" event={"ID":"da03c2c2-f43a-4f51-9c87-8e2f6b0af367","Type":"ContainerDied","Data":"8e9ebdba2b7ba70e002116caf37288c6ad9809835e0e3e9cd9c81dc7b7cf0d7f"} Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.276304 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.295256 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ceph\") pod \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.296671 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ssh-key\") pod \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.296768 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr6jv\" (UniqueName: \"kubernetes.io/projected/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-kube-api-access-mr6jv\") pod \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.296800 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-inventory\") pod \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\" (UID: \"da03c2c2-f43a-4f51-9c87-8e2f6b0af367\") " Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.302015 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-kube-api-access-mr6jv" (OuterVolumeSpecName: "kube-api-access-mr6jv") pod "da03c2c2-f43a-4f51-9c87-8e2f6b0af367" (UID: "da03c2c2-f43a-4f51-9c87-8e2f6b0af367"). InnerVolumeSpecName "kube-api-access-mr6jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.302244 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ceph" (OuterVolumeSpecName: "ceph") pod "da03c2c2-f43a-4f51-9c87-8e2f6b0af367" (UID: "da03c2c2-f43a-4f51-9c87-8e2f6b0af367"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.324390 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "da03c2c2-f43a-4f51-9c87-8e2f6b0af367" (UID: "da03c2c2-f43a-4f51-9c87-8e2f6b0af367"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.342230 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-inventory" (OuterVolumeSpecName: "inventory") pod "da03c2c2-f43a-4f51-9c87-8e2f6b0af367" (UID: "da03c2c2-f43a-4f51-9c87-8e2f6b0af367"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.398671 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr6jv\" (UniqueName: \"kubernetes.io/projected/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-kube-api-access-mr6jv\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.398710 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.398721 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.398730 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da03c2c2-f43a-4f51-9c87-8e2f6b0af367-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.843228 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" event={"ID":"da03c2c2-f43a-4f51-9c87-8e2f6b0af367","Type":"ContainerDied","Data":"4c1586d86646b190d71b4dace567af707674ce86215b6dd8cf184158785b5f73"} Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.843551 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c1586d86646b190d71b4dace567af707674ce86215b6dd8cf184158785b5f73" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.843314 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.924117 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs"] Nov 23 20:46:34 crc kubenswrapper[4726]: E1123 20:46:34.924568 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da03c2c2-f43a-4f51-9c87-8e2f6b0af367" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.924593 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="da03c2c2-f43a-4f51-9c87-8e2f6b0af367" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.924849 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="da03c2c2-f43a-4f51-9c87-8e2f6b0af367" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.925643 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.932803 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs"] Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.935665 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.935849 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.936216 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.937835 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:46:34 crc kubenswrapper[4726]: I1123 20:46:34.938375 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.005272 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.005416 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg5bf\" (UniqueName: \"kubernetes.io/projected/541068f2-1c18-4bc7-8c86-51f768e97783-kube-api-access-hg5bf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.005511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.005612 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.107151 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.107204 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.107256 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.107334 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg5bf\" (UniqueName: \"kubernetes.io/projected/541068f2-1c18-4bc7-8c86-51f768e97783-kube-api-access-hg5bf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.111050 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.111141 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.111708 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.123328 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg5bf\" (UniqueName: \"kubernetes.io/projected/541068f2-1c18-4bc7-8c86-51f768e97783-kube-api-access-hg5bf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-972gs\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.245339 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.607098 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs"] Nov 23 20:46:35 crc kubenswrapper[4726]: I1123 20:46:35.875075 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" event={"ID":"541068f2-1c18-4bc7-8c86-51f768e97783","Type":"ContainerStarted","Data":"c0f2e1b9cf94e440744e385d5d51115b1434c9fd8e1dfd09ea9ca4117a5a09f1"} Nov 23 20:46:36 crc kubenswrapper[4726]: I1123 20:46:36.884649 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" event={"ID":"541068f2-1c18-4bc7-8c86-51f768e97783","Type":"ContainerStarted","Data":"4a538cdc0a199cc93727f84ad94cceaac6b1e001245f7902e801fa206737acff"} Nov 23 20:46:36 crc kubenswrapper[4726]: I1123 20:46:36.910606 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" podStartSLOduration=2.5018009020000003 podStartE2EDuration="2.910588118s" podCreationTimestamp="2025-11-23 20:46:34 +0000 UTC" firstStartedPulling="2025-11-23 20:46:35.622347304 +0000 UTC m=+2303.771388260" lastFinishedPulling="2025-11-23 20:46:36.03113452 +0000 UTC m=+2304.180175476" observedRunningTime="2025-11-23 20:46:36.903289601 +0000 UTC m=+2305.052330557" watchObservedRunningTime="2025-11-23 20:46:36.910588118 +0000 UTC m=+2305.059629064" Nov 23 20:46:40 crc kubenswrapper[4726]: I1123 20:46:40.588860 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:46:40 crc kubenswrapper[4726]: E1123 20:46:40.594147 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:46:53 crc kubenswrapper[4726]: I1123 20:46:53.589223 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:46:53 crc kubenswrapper[4726]: E1123 20:46:53.590655 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:47:05 crc kubenswrapper[4726]: I1123 20:47:05.589421 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:47:05 crc kubenswrapper[4726]: E1123 20:47:05.590276 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:47:17 crc kubenswrapper[4726]: I1123 20:47:17.588952 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:47:17 crc kubenswrapper[4726]: E1123 20:47:17.589818 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:47:19 crc kubenswrapper[4726]: I1123 20:47:19.260045 4726 generic.go:334] "Generic (PLEG): container finished" podID="541068f2-1c18-4bc7-8c86-51f768e97783" containerID="4a538cdc0a199cc93727f84ad94cceaac6b1e001245f7902e801fa206737acff" exitCode=0 Nov 23 20:47:19 crc kubenswrapper[4726]: I1123 20:47:19.260151 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" event={"ID":"541068f2-1c18-4bc7-8c86-51f768e97783","Type":"ContainerDied","Data":"4a538cdc0a199cc93727f84ad94cceaac6b1e001245f7902e801fa206737acff"} Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.738860 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.889222 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-inventory\") pod \"541068f2-1c18-4bc7-8c86-51f768e97783\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.889279 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ceph\") pod \"541068f2-1c18-4bc7-8c86-51f768e97783\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.889340 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg5bf\" (UniqueName: \"kubernetes.io/projected/541068f2-1c18-4bc7-8c86-51f768e97783-kube-api-access-hg5bf\") pod \"541068f2-1c18-4bc7-8c86-51f768e97783\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.889463 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ssh-key\") pod \"541068f2-1c18-4bc7-8c86-51f768e97783\" (UID: \"541068f2-1c18-4bc7-8c86-51f768e97783\") " Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.894625 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/541068f2-1c18-4bc7-8c86-51f768e97783-kube-api-access-hg5bf" (OuterVolumeSpecName: "kube-api-access-hg5bf") pod "541068f2-1c18-4bc7-8c86-51f768e97783" (UID: "541068f2-1c18-4bc7-8c86-51f768e97783"). InnerVolumeSpecName "kube-api-access-hg5bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.898039 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ceph" (OuterVolumeSpecName: "ceph") pod "541068f2-1c18-4bc7-8c86-51f768e97783" (UID: "541068f2-1c18-4bc7-8c86-51f768e97783"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.915526 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "541068f2-1c18-4bc7-8c86-51f768e97783" (UID: "541068f2-1c18-4bc7-8c86-51f768e97783"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.917647 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-inventory" (OuterVolumeSpecName: "inventory") pod "541068f2-1c18-4bc7-8c86-51f768e97783" (UID: "541068f2-1c18-4bc7-8c86-51f768e97783"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.991621 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.991655 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.991665 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/541068f2-1c18-4bc7-8c86-51f768e97783-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:47:20 crc kubenswrapper[4726]: I1123 20:47:20.991674 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg5bf\" (UniqueName: \"kubernetes.io/projected/541068f2-1c18-4bc7-8c86-51f768e97783-kube-api-access-hg5bf\") on node \"crc\" DevicePath \"\"" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.289127 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" event={"ID":"541068f2-1c18-4bc7-8c86-51f768e97783","Type":"ContainerDied","Data":"c0f2e1b9cf94e440744e385d5d51115b1434c9fd8e1dfd09ea9ca4117a5a09f1"} Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.289172 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0f2e1b9cf94e440744e385d5d51115b1434c9fd8e1dfd09ea9ca4117a5a09f1" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.289424 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-972gs" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.400210 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2"] Nov 23 20:47:21 crc kubenswrapper[4726]: E1123 20:47:21.400969 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="541068f2-1c18-4bc7-8c86-51f768e97783" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.400990 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="541068f2-1c18-4bc7-8c86-51f768e97783" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.401245 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="541068f2-1c18-4bc7-8c86-51f768e97783" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.402081 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.408776 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2"] Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.450590 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.450668 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.450832 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.451021 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.451136 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.504296 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.504594 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgkqp\" (UniqueName: \"kubernetes.io/projected/0c1ed863-f501-4aec-bafe-2d7d828b8d78-kube-api-access-cgkqp\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.504765 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.504936 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.606302 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.606604 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgkqp\" (UniqueName: \"kubernetes.io/projected/0c1ed863-f501-4aec-bafe-2d7d828b8d78-kube-api-access-cgkqp\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.606735 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.607115 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.618660 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.618700 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.620347 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.626953 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgkqp\" (UniqueName: \"kubernetes.io/projected/0c1ed863-f501-4aec-bafe-2d7d828b8d78-kube-api-access-cgkqp\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:21 crc kubenswrapper[4726]: I1123 20:47:21.780074 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:22 crc kubenswrapper[4726]: I1123 20:47:22.292928 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2"] Nov 23 20:47:23 crc kubenswrapper[4726]: I1123 20:47:23.309667 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" event={"ID":"0c1ed863-f501-4aec-bafe-2d7d828b8d78","Type":"ContainerStarted","Data":"9bbb62f1738687a41cce6fa05a5b1985c0d2693d109dcd1f62cf635c4bb817b0"} Nov 23 20:47:23 crc kubenswrapper[4726]: I1123 20:47:23.310131 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" event={"ID":"0c1ed863-f501-4aec-bafe-2d7d828b8d78","Type":"ContainerStarted","Data":"59e9798621f7565253ddb6da240ac1f3beef73ef8cff7383d686a260dd85726a"} Nov 23 20:47:23 crc kubenswrapper[4726]: I1123 20:47:23.348282 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" podStartSLOduration=1.940238827 podStartE2EDuration="2.348252943s" podCreationTimestamp="2025-11-23 20:47:21 +0000 UTC" firstStartedPulling="2025-11-23 20:47:22.305672625 +0000 UTC m=+2350.454713581" lastFinishedPulling="2025-11-23 20:47:22.713686701 +0000 UTC m=+2350.862727697" observedRunningTime="2025-11-23 20:47:23.333118725 +0000 UTC m=+2351.482159681" watchObservedRunningTime="2025-11-23 20:47:23.348252943 +0000 UTC m=+2351.497293929" Nov 23 20:47:27 crc kubenswrapper[4726]: I1123 20:47:27.346748 4726 generic.go:334] "Generic (PLEG): container finished" podID="0c1ed863-f501-4aec-bafe-2d7d828b8d78" containerID="9bbb62f1738687a41cce6fa05a5b1985c0d2693d109dcd1f62cf635c4bb817b0" exitCode=0 Nov 23 20:47:27 crc kubenswrapper[4726]: I1123 20:47:27.346859 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" event={"ID":"0c1ed863-f501-4aec-bafe-2d7d828b8d78","Type":"ContainerDied","Data":"9bbb62f1738687a41cce6fa05a5b1985c0d2693d109dcd1f62cf635c4bb817b0"} Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.797011 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.835913 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ceph\") pod \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.836004 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ssh-key\") pod \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.836213 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgkqp\" (UniqueName: \"kubernetes.io/projected/0c1ed863-f501-4aec-bafe-2d7d828b8d78-kube-api-access-cgkqp\") pod \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.836297 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-inventory\") pod \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\" (UID: \"0c1ed863-f501-4aec-bafe-2d7d828b8d78\") " Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.844069 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ceph" (OuterVolumeSpecName: "ceph") pod "0c1ed863-f501-4aec-bafe-2d7d828b8d78" (UID: "0c1ed863-f501-4aec-bafe-2d7d828b8d78"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.844575 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c1ed863-f501-4aec-bafe-2d7d828b8d78-kube-api-access-cgkqp" (OuterVolumeSpecName: "kube-api-access-cgkqp") pod "0c1ed863-f501-4aec-bafe-2d7d828b8d78" (UID: "0c1ed863-f501-4aec-bafe-2d7d828b8d78"). InnerVolumeSpecName "kube-api-access-cgkqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.874755 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0c1ed863-f501-4aec-bafe-2d7d828b8d78" (UID: "0c1ed863-f501-4aec-bafe-2d7d828b8d78"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.880092 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-inventory" (OuterVolumeSpecName: "inventory") pod "0c1ed863-f501-4aec-bafe-2d7d828b8d78" (UID: "0c1ed863-f501-4aec-bafe-2d7d828b8d78"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.938304 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgkqp\" (UniqueName: \"kubernetes.io/projected/0c1ed863-f501-4aec-bafe-2d7d828b8d78-kube-api-access-cgkqp\") on node \"crc\" DevicePath \"\"" Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.938350 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.938362 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:47:28 crc kubenswrapper[4726]: I1123 20:47:28.938373 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c1ed863-f501-4aec-bafe-2d7d828b8d78-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.369369 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" event={"ID":"0c1ed863-f501-4aec-bafe-2d7d828b8d78","Type":"ContainerDied","Data":"59e9798621f7565253ddb6da240ac1f3beef73ef8cff7383d686a260dd85726a"} Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.369408 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59e9798621f7565253ddb6da240ac1f3beef73ef8cff7383d686a260dd85726a" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.369495 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.471061 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr"] Nov 23 20:47:29 crc kubenswrapper[4726]: E1123 20:47:29.471853 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1ed863-f501-4aec-bafe-2d7d828b8d78" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.471920 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1ed863-f501-4aec-bafe-2d7d828b8d78" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.472187 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1ed863-f501-4aec-bafe-2d7d828b8d78" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.473553 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.476257 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.476315 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.477768 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.481793 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr"] Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.482813 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.482818 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.550709 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.550791 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6dhz\" (UniqueName: \"kubernetes.io/projected/77123683-ed45-435f-b733-f5cc3822169c-kube-api-access-q6dhz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.550840 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.550889 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.652339 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.652427 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6dhz\" (UniqueName: \"kubernetes.io/projected/77123683-ed45-435f-b733-f5cc3822169c-kube-api-access-q6dhz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.652455 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.652480 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.658015 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.659723 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.668295 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.676610 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6dhz\" (UniqueName: \"kubernetes.io/projected/77123683-ed45-435f-b733-f5cc3822169c-kube-api-access-q6dhz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:29 crc kubenswrapper[4726]: I1123 20:47:29.808092 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:47:30 crc kubenswrapper[4726]: I1123 20:47:30.157793 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr"] Nov 23 20:47:30 crc kubenswrapper[4726]: I1123 20:47:30.377897 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" event={"ID":"77123683-ed45-435f-b733-f5cc3822169c","Type":"ContainerStarted","Data":"3c5d54907b2e6429198dd35504bb1d009ee26c4223ee701744226ee47b6895cf"} Nov 23 20:47:30 crc kubenswrapper[4726]: I1123 20:47:30.589502 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:47:30 crc kubenswrapper[4726]: E1123 20:47:30.590316 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:47:31 crc kubenswrapper[4726]: I1123 20:47:31.387729 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" event={"ID":"77123683-ed45-435f-b733-f5cc3822169c","Type":"ContainerStarted","Data":"ca5b46d9f9793daa95ec748748a13317151bf5fe21291ab49f556c68774ea95a"} Nov 23 20:47:41 crc kubenswrapper[4726]: I1123 20:47:41.589193 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:47:41 crc kubenswrapper[4726]: E1123 20:47:41.589957 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:47:52 crc kubenswrapper[4726]: I1123 20:47:52.596288 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:47:52 crc kubenswrapper[4726]: E1123 20:47:52.597305 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:48:06 crc kubenswrapper[4726]: I1123 20:48:06.588960 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:48:06 crc kubenswrapper[4726]: E1123 20:48:06.590016 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:48:21 crc kubenswrapper[4726]: I1123 20:48:21.589741 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:48:21 crc kubenswrapper[4726]: E1123 20:48:21.591546 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:48:23 crc kubenswrapper[4726]: I1123 20:48:23.870314 4726 generic.go:334] "Generic (PLEG): container finished" podID="77123683-ed45-435f-b733-f5cc3822169c" containerID="ca5b46d9f9793daa95ec748748a13317151bf5fe21291ab49f556c68774ea95a" exitCode=0 Nov 23 20:48:23 crc kubenswrapper[4726]: I1123 20:48:23.870427 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" event={"ID":"77123683-ed45-435f-b733-f5cc3822169c","Type":"ContainerDied","Data":"ca5b46d9f9793daa95ec748748a13317151bf5fe21291ab49f556c68774ea95a"} Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.468812 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.568390 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6dhz\" (UniqueName: \"kubernetes.io/projected/77123683-ed45-435f-b733-f5cc3822169c-kube-api-access-q6dhz\") pod \"77123683-ed45-435f-b733-f5cc3822169c\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.568658 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ceph\") pod \"77123683-ed45-435f-b733-f5cc3822169c\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.568878 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-inventory\") pod \"77123683-ed45-435f-b733-f5cc3822169c\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.569027 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ssh-key\") pod \"77123683-ed45-435f-b733-f5cc3822169c\" (UID: \"77123683-ed45-435f-b733-f5cc3822169c\") " Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.574132 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ceph" (OuterVolumeSpecName: "ceph") pod "77123683-ed45-435f-b733-f5cc3822169c" (UID: "77123683-ed45-435f-b733-f5cc3822169c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.575689 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77123683-ed45-435f-b733-f5cc3822169c-kube-api-access-q6dhz" (OuterVolumeSpecName: "kube-api-access-q6dhz") pod "77123683-ed45-435f-b733-f5cc3822169c" (UID: "77123683-ed45-435f-b733-f5cc3822169c"). InnerVolumeSpecName "kube-api-access-q6dhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.599165 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "77123683-ed45-435f-b733-f5cc3822169c" (UID: "77123683-ed45-435f-b733-f5cc3822169c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.606002 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-inventory" (OuterVolumeSpecName: "inventory") pod "77123683-ed45-435f-b733-f5cc3822169c" (UID: "77123683-ed45-435f-b733-f5cc3822169c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.672413 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6dhz\" (UniqueName: \"kubernetes.io/projected/77123683-ed45-435f-b733-f5cc3822169c-kube-api-access-q6dhz\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.672439 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.672449 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.672459 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77123683-ed45-435f-b733-f5cc3822169c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.893799 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" event={"ID":"77123683-ed45-435f-b733-f5cc3822169c","Type":"ContainerDied","Data":"3c5d54907b2e6429198dd35504bb1d009ee26c4223ee701744226ee47b6895cf"} Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.894165 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c5d54907b2e6429198dd35504bb1d009ee26c4223ee701744226ee47b6895cf" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.893960 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr" Nov 23 20:48:25 crc kubenswrapper[4726]: I1123 20:48:25.999821 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mvtwf"] Nov 23 20:48:26 crc kubenswrapper[4726]: E1123 20:48:26.000337 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77123683-ed45-435f-b733-f5cc3822169c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.000427 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="77123683-ed45-435f-b733-f5cc3822169c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.000677 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="77123683-ed45-435f-b733-f5cc3822169c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.001416 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.003661 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.004806 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.005137 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.006073 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.009581 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.028072 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mvtwf"] Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.079221 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ceph\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.079287 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dqpl\" (UniqueName: \"kubernetes.io/projected/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-kube-api-access-9dqpl\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.079321 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.079759 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.181919 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.181995 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ceph\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.182026 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dqpl\" (UniqueName: \"kubernetes.io/projected/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-kube-api-access-9dqpl\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.182047 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.189320 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.189605 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ceph\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.206935 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dqpl\" (UniqueName: \"kubernetes.io/projected/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-kube-api-access-9dqpl\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.209794 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mvtwf\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.317501 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.856486 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mvtwf"] Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.866607 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 20:48:26 crc kubenswrapper[4726]: I1123 20:48:26.902293 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" event={"ID":"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8","Type":"ContainerStarted","Data":"1ea010ac914f57e39e289b1539d29a922e7c3a5ed64fdfaacc061028881b90d5"} Nov 23 20:48:27 crc kubenswrapper[4726]: I1123 20:48:27.913256 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" event={"ID":"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8","Type":"ContainerStarted","Data":"7a2c756b58f04c2d935785eea94ff60fc3b6cb573e89d92746e3c7b0223dc0bc"} Nov 23 20:48:27 crc kubenswrapper[4726]: I1123 20:48:27.949549 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" podStartSLOduration=2.4943639 podStartE2EDuration="2.949525889s" podCreationTimestamp="2025-11-23 20:48:25 +0000 UTC" firstStartedPulling="2025-11-23 20:48:26.866417995 +0000 UTC m=+2415.015458951" lastFinishedPulling="2025-11-23 20:48:27.321579984 +0000 UTC m=+2415.470620940" observedRunningTime="2025-11-23 20:48:27.943317613 +0000 UTC m=+2416.092358589" watchObservedRunningTime="2025-11-23 20:48:27.949525889 +0000 UTC m=+2416.098566845" Nov 23 20:48:34 crc kubenswrapper[4726]: I1123 20:48:34.589650 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:48:34 crc kubenswrapper[4726]: E1123 20:48:34.591437 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:48:39 crc kubenswrapper[4726]: I1123 20:48:39.006948 4726 generic.go:334] "Generic (PLEG): container finished" podID="80b8ce4c-915f-424b-99c4-a16f0d9a1fc8" containerID="7a2c756b58f04c2d935785eea94ff60fc3b6cb573e89d92746e3c7b0223dc0bc" exitCode=0 Nov 23 20:48:39 crc kubenswrapper[4726]: I1123 20:48:39.007036 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" event={"ID":"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8","Type":"ContainerDied","Data":"7a2c756b58f04c2d935785eea94ff60fc3b6cb573e89d92746e3c7b0223dc0bc"} Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.417962 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.466816 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-inventory-0\") pod \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.467214 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ceph\") pod \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.467495 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dqpl\" (UniqueName: \"kubernetes.io/projected/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-kube-api-access-9dqpl\") pod \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.467715 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ssh-key-openstack-edpm-ipam\") pod \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\" (UID: \"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8\") " Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.474234 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-kube-api-access-9dqpl" (OuterVolumeSpecName: "kube-api-access-9dqpl") pod "80b8ce4c-915f-424b-99c4-a16f0d9a1fc8" (UID: "80b8ce4c-915f-424b-99c4-a16f0d9a1fc8"). InnerVolumeSpecName "kube-api-access-9dqpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.476066 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ceph" (OuterVolumeSpecName: "ceph") pod "80b8ce4c-915f-424b-99c4-a16f0d9a1fc8" (UID: "80b8ce4c-915f-424b-99c4-a16f0d9a1fc8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.491629 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "80b8ce4c-915f-424b-99c4-a16f0d9a1fc8" (UID: "80b8ce4c-915f-424b-99c4-a16f0d9a1fc8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.501422 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "80b8ce4c-915f-424b-99c4-a16f0d9a1fc8" (UID: "80b8ce4c-915f-424b-99c4-a16f0d9a1fc8"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.570043 4726 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.570076 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.570086 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dqpl\" (UniqueName: \"kubernetes.io/projected/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-kube-api-access-9dqpl\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:40 crc kubenswrapper[4726]: I1123 20:48:40.570097 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80b8ce4c-915f-424b-99c4-a16f0d9a1fc8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.040742 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" event={"ID":"80b8ce4c-915f-424b-99c4-a16f0d9a1fc8","Type":"ContainerDied","Data":"1ea010ac914f57e39e289b1539d29a922e7c3a5ed64fdfaacc061028881b90d5"} Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.040793 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ea010ac914f57e39e289b1539d29a922e7c3a5ed64fdfaacc061028881b90d5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.040814 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mvtwf" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.146088 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5"] Nov 23 20:48:41 crc kubenswrapper[4726]: E1123 20:48:41.147222 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b8ce4c-915f-424b-99c4-a16f0d9a1fc8" containerName="ssh-known-hosts-edpm-deployment" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.147248 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b8ce4c-915f-424b-99c4-a16f0d9a1fc8" containerName="ssh-known-hosts-edpm-deployment" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.147576 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b8ce4c-915f-424b-99c4-a16f0d9a1fc8" containerName="ssh-known-hosts-edpm-deployment" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.149043 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.151817 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.153134 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.153279 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.153359 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.154184 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.163706 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5"] Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.180249 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.180325 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.180351 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.180525 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-878dl\" (UniqueName: \"kubernetes.io/projected/275bc578-25bb-4340-9e5c-f2c296713110-kube-api-access-878dl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.281955 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.282003 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.282046 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-878dl\" (UniqueName: \"kubernetes.io/projected/275bc578-25bb-4340-9e5c-f2c296713110-kube-api-access-878dl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.282178 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.286952 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.287371 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.290448 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.304194 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-878dl\" (UniqueName: \"kubernetes.io/projected/275bc578-25bb-4340-9e5c-f2c296713110-kube-api-access-878dl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pvpj5\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.465253 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:41 crc kubenswrapper[4726]: I1123 20:48:41.994760 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5"] Nov 23 20:48:42 crc kubenswrapper[4726]: I1123 20:48:42.051746 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" event={"ID":"275bc578-25bb-4340-9e5c-f2c296713110","Type":"ContainerStarted","Data":"85b1f7b4e5db5f3eb80580e19a9d15fe246f5d93ded1b25a6217f41f8d60018c"} Nov 23 20:48:43 crc kubenswrapper[4726]: I1123 20:48:43.065928 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" event={"ID":"275bc578-25bb-4340-9e5c-f2c296713110","Type":"ContainerStarted","Data":"7deb0135456501ae0549499741012f9fa4a6c5990c56d69a72fc0e8fe116f855"} Nov 23 20:48:43 crc kubenswrapper[4726]: I1123 20:48:43.092856 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" podStartSLOduration=1.609136966 podStartE2EDuration="2.092828642s" podCreationTimestamp="2025-11-23 20:48:41 +0000 UTC" firstStartedPulling="2025-11-23 20:48:42.010058097 +0000 UTC m=+2430.159099083" lastFinishedPulling="2025-11-23 20:48:42.493749803 +0000 UTC m=+2430.642790759" observedRunningTime="2025-11-23 20:48:43.088574512 +0000 UTC m=+2431.237615548" watchObservedRunningTime="2025-11-23 20:48:43.092828642 +0000 UTC m=+2431.241869628" Nov 23 20:48:48 crc kubenswrapper[4726]: I1123 20:48:48.590169 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:48:48 crc kubenswrapper[4726]: E1123 20:48:48.591457 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:48:52 crc kubenswrapper[4726]: I1123 20:48:52.144661 4726 generic.go:334] "Generic (PLEG): container finished" podID="275bc578-25bb-4340-9e5c-f2c296713110" containerID="7deb0135456501ae0549499741012f9fa4a6c5990c56d69a72fc0e8fe116f855" exitCode=0 Nov 23 20:48:52 crc kubenswrapper[4726]: I1123 20:48:52.144711 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" event={"ID":"275bc578-25bb-4340-9e5c-f2c296713110","Type":"ContainerDied","Data":"7deb0135456501ae0549499741012f9fa4a6c5990c56d69a72fc0e8fe116f855"} Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.597352 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.782735 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ceph\") pod \"275bc578-25bb-4340-9e5c-f2c296713110\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.782807 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ssh-key\") pod \"275bc578-25bb-4340-9e5c-f2c296713110\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.783058 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-878dl\" (UniqueName: \"kubernetes.io/projected/275bc578-25bb-4340-9e5c-f2c296713110-kube-api-access-878dl\") pod \"275bc578-25bb-4340-9e5c-f2c296713110\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.783281 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-inventory\") pod \"275bc578-25bb-4340-9e5c-f2c296713110\" (UID: \"275bc578-25bb-4340-9e5c-f2c296713110\") " Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.791473 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/275bc578-25bb-4340-9e5c-f2c296713110-kube-api-access-878dl" (OuterVolumeSpecName: "kube-api-access-878dl") pod "275bc578-25bb-4340-9e5c-f2c296713110" (UID: "275bc578-25bb-4340-9e5c-f2c296713110"). InnerVolumeSpecName "kube-api-access-878dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.806559 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ceph" (OuterVolumeSpecName: "ceph") pod "275bc578-25bb-4340-9e5c-f2c296713110" (UID: "275bc578-25bb-4340-9e5c-f2c296713110"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.814990 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "275bc578-25bb-4340-9e5c-f2c296713110" (UID: "275bc578-25bb-4340-9e5c-f2c296713110"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.820142 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-inventory" (OuterVolumeSpecName: "inventory") pod "275bc578-25bb-4340-9e5c-f2c296713110" (UID: "275bc578-25bb-4340-9e5c-f2c296713110"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.885177 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.885526 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.885537 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-878dl\" (UniqueName: \"kubernetes.io/projected/275bc578-25bb-4340-9e5c-f2c296713110-kube-api-access-878dl\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:53 crc kubenswrapper[4726]: I1123 20:48:53.885546 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/275bc578-25bb-4340-9e5c-f2c296713110-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.165227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" event={"ID":"275bc578-25bb-4340-9e5c-f2c296713110","Type":"ContainerDied","Data":"85b1f7b4e5db5f3eb80580e19a9d15fe246f5d93ded1b25a6217f41f8d60018c"} Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.165281 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85b1f7b4e5db5f3eb80580e19a9d15fe246f5d93ded1b25a6217f41f8d60018c" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.165440 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pvpj5" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.240641 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6"] Nov 23 20:48:54 crc kubenswrapper[4726]: E1123 20:48:54.241021 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275bc578-25bb-4340-9e5c-f2c296713110" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.241039 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="275bc578-25bb-4340-9e5c-f2c296713110" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.241220 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="275bc578-25bb-4340-9e5c-f2c296713110" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.241844 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.243543 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.244343 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.244542 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.244754 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.246917 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.263371 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6"] Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.292930 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.292991 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdv9p\" (UniqueName: \"kubernetes.io/projected/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-kube-api-access-vdv9p\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.293059 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.293088 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.394536 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.394694 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.394765 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdv9p\" (UniqueName: \"kubernetes.io/projected/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-kube-api-access-vdv9p\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.394927 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.398781 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.398898 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.405375 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.416521 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdv9p\" (UniqueName: \"kubernetes.io/projected/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-kube-api-access-vdv9p\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:54 crc kubenswrapper[4726]: I1123 20:48:54.558063 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:48:55 crc kubenswrapper[4726]: I1123 20:48:55.098713 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6"] Nov 23 20:48:55 crc kubenswrapper[4726]: I1123 20:48:55.172560 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" event={"ID":"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4","Type":"ContainerStarted","Data":"1564ac38c00e9388210807c6392a12eaad11b9ac349017aaa14ba8fe2569f7ef"} Nov 23 20:48:56 crc kubenswrapper[4726]: I1123 20:48:56.181254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" event={"ID":"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4","Type":"ContainerStarted","Data":"d0e9d0d3b3ff61b9ddceaa033ad4804147cc18a9174377af80bd56ceac7975b9"} Nov 23 20:48:56 crc kubenswrapper[4726]: I1123 20:48:56.210232 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" podStartSLOduration=1.73053163 podStartE2EDuration="2.210204222s" podCreationTimestamp="2025-11-23 20:48:54 +0000 UTC" firstStartedPulling="2025-11-23 20:48:55.121117269 +0000 UTC m=+2443.270158225" lastFinishedPulling="2025-11-23 20:48:55.600789831 +0000 UTC m=+2443.749830817" observedRunningTime="2025-11-23 20:48:56.19952906 +0000 UTC m=+2444.348570076" watchObservedRunningTime="2025-11-23 20:48:56.210204222 +0000 UTC m=+2444.359245188" Nov 23 20:49:00 crc kubenswrapper[4726]: I1123 20:49:00.589241 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:49:00 crc kubenswrapper[4726]: E1123 20:49:00.591984 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:49:06 crc kubenswrapper[4726]: I1123 20:49:06.259172 4726 generic.go:334] "Generic (PLEG): container finished" podID="e1af3b69-28d2-4a52-8f13-7cb3e9831fb4" containerID="d0e9d0d3b3ff61b9ddceaa033ad4804147cc18a9174377af80bd56ceac7975b9" exitCode=0 Nov 23 20:49:06 crc kubenswrapper[4726]: I1123 20:49:06.259224 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" event={"ID":"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4","Type":"ContainerDied","Data":"d0e9d0d3b3ff61b9ddceaa033ad4804147cc18a9174377af80bd56ceac7975b9"} Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.694664 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.757626 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ceph\") pod \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.757936 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ssh-key\") pod \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.758685 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-inventory\") pod \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.758978 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdv9p\" (UniqueName: \"kubernetes.io/projected/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-kube-api-access-vdv9p\") pod \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\" (UID: \"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4\") " Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.763248 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-kube-api-access-vdv9p" (OuterVolumeSpecName: "kube-api-access-vdv9p") pod "e1af3b69-28d2-4a52-8f13-7cb3e9831fb4" (UID: "e1af3b69-28d2-4a52-8f13-7cb3e9831fb4"). InnerVolumeSpecName "kube-api-access-vdv9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.763969 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ceph" (OuterVolumeSpecName: "ceph") pod "e1af3b69-28d2-4a52-8f13-7cb3e9831fb4" (UID: "e1af3b69-28d2-4a52-8f13-7cb3e9831fb4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.791003 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-inventory" (OuterVolumeSpecName: "inventory") pod "e1af3b69-28d2-4a52-8f13-7cb3e9831fb4" (UID: "e1af3b69-28d2-4a52-8f13-7cb3e9831fb4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.794024 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e1af3b69-28d2-4a52-8f13-7cb3e9831fb4" (UID: "e1af3b69-28d2-4a52-8f13-7cb3e9831fb4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.862665 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdv9p\" (UniqueName: \"kubernetes.io/projected/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-kube-api-access-vdv9p\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.862688 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.863687 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:07 crc kubenswrapper[4726]: I1123 20:49:07.864337 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1af3b69-28d2-4a52-8f13-7cb3e9831fb4-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.288470 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" event={"ID":"e1af3b69-28d2-4a52-8f13-7cb3e9831fb4","Type":"ContainerDied","Data":"1564ac38c00e9388210807c6392a12eaad11b9ac349017aaa14ba8fe2569f7ef"} Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.288518 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1564ac38c00e9388210807c6392a12eaad11b9ac349017aaa14ba8fe2569f7ef" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.288628 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.457011 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d"] Nov 23 20:49:08 crc kubenswrapper[4726]: E1123 20:49:08.457533 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1af3b69-28d2-4a52-8f13-7cb3e9831fb4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.457552 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1af3b69-28d2-4a52-8f13-7cb3e9831fb4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.457822 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1af3b69-28d2-4a52-8f13-7cb3e9831fb4" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.458674 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.463023 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d"] Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.463575 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.463947 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.464233 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.464762 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.464968 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.465098 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.465163 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.465093 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577298 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577342 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577375 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577479 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577563 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577664 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577721 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577750 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577768 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4sv7\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-kube-api-access-r4sv7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577797 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577859 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.577963 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.578005 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680282 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680377 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680424 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680451 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680475 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4sv7\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-kube-api-access-r4sv7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680519 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680576 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680638 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680673 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680747 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680772 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680807 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.680835 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.686744 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.686814 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.689517 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.690502 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.695368 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.695668 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.697594 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.712450 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.713000 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.714552 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.715265 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.715700 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4sv7\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-kube-api-access-r4sv7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.721362 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:08 crc kubenswrapper[4726]: I1123 20:49:08.816013 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:09 crc kubenswrapper[4726]: I1123 20:49:09.332598 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d"] Nov 23 20:49:10 crc kubenswrapper[4726]: I1123 20:49:10.307586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" event={"ID":"01fcd196-1da1-4fb2-a900-f3960109b195","Type":"ContainerStarted","Data":"931743e13d2d84b169dd36eae25d737ff3d6a0a1b4f1e1b4b28752b6fef7e0fd"} Nov 23 20:49:10 crc kubenswrapper[4726]: I1123 20:49:10.308916 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" event={"ID":"01fcd196-1da1-4fb2-a900-f3960109b195","Type":"ContainerStarted","Data":"50d646b839b85a74716365da67749fcff12b2f2537955ad329481d0a15dc7c78"} Nov 23 20:49:10 crc kubenswrapper[4726]: I1123 20:49:10.355222 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" podStartSLOduration=1.935205373 podStartE2EDuration="2.355204998s" podCreationTimestamp="2025-11-23 20:49:08 +0000 UTC" firstStartedPulling="2025-11-23 20:49:09.336565457 +0000 UTC m=+2457.485606413" lastFinishedPulling="2025-11-23 20:49:09.756565072 +0000 UTC m=+2457.905606038" observedRunningTime="2025-11-23 20:49:10.349728643 +0000 UTC m=+2458.498769599" watchObservedRunningTime="2025-11-23 20:49:10.355204998 +0000 UTC m=+2458.504245954" Nov 23 20:49:12 crc kubenswrapper[4726]: I1123 20:49:12.595935 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:49:12 crc kubenswrapper[4726]: E1123 20:49:12.596490 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:49:23 crc kubenswrapper[4726]: I1123 20:49:23.596880 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:49:23 crc kubenswrapper[4726]: E1123 20:49:23.597737 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:49:34 crc kubenswrapper[4726]: I1123 20:49:34.588853 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:49:34 crc kubenswrapper[4726]: E1123 20:49:34.589629 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:49:45 crc kubenswrapper[4726]: I1123 20:49:45.589681 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:49:46 crc kubenswrapper[4726]: I1123 20:49:46.692921 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"5f7ac676ac424776f668164058671554206e267544abf2ca6f34ada9645d7798"} Nov 23 20:49:47 crc kubenswrapper[4726]: I1123 20:49:47.710679 4726 generic.go:334] "Generic (PLEG): container finished" podID="01fcd196-1da1-4fb2-a900-f3960109b195" containerID="931743e13d2d84b169dd36eae25d737ff3d6a0a1b4f1e1b4b28752b6fef7e0fd" exitCode=0 Nov 23 20:49:47 crc kubenswrapper[4726]: I1123 20:49:47.710789 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" event={"ID":"01fcd196-1da1-4fb2-a900-f3960109b195","Type":"ContainerDied","Data":"931743e13d2d84b169dd36eae25d737ff3d6a0a1b4f1e1b4b28752b6fef7e0fd"} Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.125413 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298219 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298269 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4sv7\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-kube-api-access-r4sv7\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ovn-combined-ca-bundle\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298339 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-libvirt-combined-ca-bundle\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298399 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298419 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-bootstrap-combined-ca-bundle\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298441 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-nova-combined-ca-bundle\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298462 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ssh-key\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298496 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-neutron-metadata-combined-ca-bundle\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298537 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ceph\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298588 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-repo-setup-combined-ca-bundle\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298611 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-inventory\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.298660 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-ovn-default-certs-0\") pod \"01fcd196-1da1-4fb2-a900-f3960109b195\" (UID: \"01fcd196-1da1-4fb2-a900-f3960109b195\") " Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.305175 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-kube-api-access-r4sv7" (OuterVolumeSpecName: "kube-api-access-r4sv7") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "kube-api-access-r4sv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.309727 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.310238 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.310316 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.310424 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.310526 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.310620 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.310610 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ceph" (OuterVolumeSpecName: "ceph") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.310705 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.310954 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.318010 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.329987 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.336206 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-inventory" (OuterVolumeSpecName: "inventory") pod "01fcd196-1da1-4fb2-a900-f3960109b195" (UID: "01fcd196-1da1-4fb2-a900-f3960109b195"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400813 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400849 4726 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400880 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400893 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400908 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400921 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4sv7\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-kube-api-access-r4sv7\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400933 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400943 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400954 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/01fcd196-1da1-4fb2-a900-f3960109b195-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400968 4726 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400980 4726 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.400989 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.401000 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01fcd196-1da1-4fb2-a900-f3960109b195-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.727430 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" event={"ID":"01fcd196-1da1-4fb2-a900-f3960109b195","Type":"ContainerDied","Data":"50d646b839b85a74716365da67749fcff12b2f2537955ad329481d0a15dc7c78"} Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.727714 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50d646b839b85a74716365da67749fcff12b2f2537955ad329481d0a15dc7c78" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.727481 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.833453 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h"] Nov 23 20:49:49 crc kubenswrapper[4726]: E1123 20:49:49.833799 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01fcd196-1da1-4fb2-a900-f3960109b195" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.833817 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="01fcd196-1da1-4fb2-a900-f3960109b195" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.834018 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="01fcd196-1da1-4fb2-a900-f3960109b195" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.834599 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.837136 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.837270 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.837316 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.838303 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.838304 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:49:49 crc kubenswrapper[4726]: I1123 20:49:49.851297 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h"] Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.010394 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.010450 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t9jf\" (UniqueName: \"kubernetes.io/projected/00e5cdd7-83a9-468b-a56f-db77fe0da375-kube-api-access-8t9jf\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.010538 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.010577 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.112202 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.112506 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t9jf\" (UniqueName: \"kubernetes.io/projected/00e5cdd7-83a9-468b-a56f-db77fe0da375-kube-api-access-8t9jf\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.112748 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.112888 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.117668 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.117822 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.118834 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.133917 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t9jf\" (UniqueName: \"kubernetes.io/projected/00e5cdd7-83a9-468b-a56f-db77fe0da375-kube-api-access-8t9jf\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.193301 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:50 crc kubenswrapper[4726]: I1123 20:49:50.825098 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h"] Nov 23 20:49:51 crc kubenswrapper[4726]: I1123 20:49:51.745568 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" event={"ID":"00e5cdd7-83a9-468b-a56f-db77fe0da375","Type":"ContainerStarted","Data":"3f1b47888de835dd508924940bc9bba9deb9a1ea97f8005af41700d655605e6c"} Nov 23 20:49:51 crc kubenswrapper[4726]: I1123 20:49:51.746059 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" event={"ID":"00e5cdd7-83a9-468b-a56f-db77fe0da375","Type":"ContainerStarted","Data":"c7a74ad6a86f92257aa00ff4affbe806c2b2dbe46762f79a5bb253b6ec3388c1"} Nov 23 20:49:51 crc kubenswrapper[4726]: I1123 20:49:51.772736 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" podStartSLOduration=2.166568929 podStartE2EDuration="2.772709278s" podCreationTimestamp="2025-11-23 20:49:49 +0000 UTC" firstStartedPulling="2025-11-23 20:49:50.839525492 +0000 UTC m=+2498.988566448" lastFinishedPulling="2025-11-23 20:49:51.445665801 +0000 UTC m=+2499.594706797" observedRunningTime="2025-11-23 20:49:51.761944704 +0000 UTC m=+2499.910985660" watchObservedRunningTime="2025-11-23 20:49:51.772709278 +0000 UTC m=+2499.921750274" Nov 23 20:49:57 crc kubenswrapper[4726]: I1123 20:49:57.799674 4726 generic.go:334] "Generic (PLEG): container finished" podID="00e5cdd7-83a9-468b-a56f-db77fe0da375" containerID="3f1b47888de835dd508924940bc9bba9deb9a1ea97f8005af41700d655605e6c" exitCode=0 Nov 23 20:49:57 crc kubenswrapper[4726]: I1123 20:49:57.799777 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" event={"ID":"00e5cdd7-83a9-468b-a56f-db77fe0da375","Type":"ContainerDied","Data":"3f1b47888de835dd508924940bc9bba9deb9a1ea97f8005af41700d655605e6c"} Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.223215 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.409984 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ceph\") pod \"00e5cdd7-83a9-468b-a56f-db77fe0da375\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.410038 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ssh-key\") pod \"00e5cdd7-83a9-468b-a56f-db77fe0da375\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.410106 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t9jf\" (UniqueName: \"kubernetes.io/projected/00e5cdd7-83a9-468b-a56f-db77fe0da375-kube-api-access-8t9jf\") pod \"00e5cdd7-83a9-468b-a56f-db77fe0da375\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.411128 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-inventory\") pod \"00e5cdd7-83a9-468b-a56f-db77fe0da375\" (UID: \"00e5cdd7-83a9-468b-a56f-db77fe0da375\") " Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.416512 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ceph" (OuterVolumeSpecName: "ceph") pod "00e5cdd7-83a9-468b-a56f-db77fe0da375" (UID: "00e5cdd7-83a9-468b-a56f-db77fe0da375"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.418048 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e5cdd7-83a9-468b-a56f-db77fe0da375-kube-api-access-8t9jf" (OuterVolumeSpecName: "kube-api-access-8t9jf") pod "00e5cdd7-83a9-468b-a56f-db77fe0da375" (UID: "00e5cdd7-83a9-468b-a56f-db77fe0da375"). InnerVolumeSpecName "kube-api-access-8t9jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.439375 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "00e5cdd7-83a9-468b-a56f-db77fe0da375" (UID: "00e5cdd7-83a9-468b-a56f-db77fe0da375"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.440669 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-inventory" (OuterVolumeSpecName: "inventory") pod "00e5cdd7-83a9-468b-a56f-db77fe0da375" (UID: "00e5cdd7-83a9-468b-a56f-db77fe0da375"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.514013 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t9jf\" (UniqueName: \"kubernetes.io/projected/00e5cdd7-83a9-468b-a56f-db77fe0da375-kube-api-access-8t9jf\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.514062 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.514081 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.515039 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00e5cdd7-83a9-468b-a56f-db77fe0da375-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.821387 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" event={"ID":"00e5cdd7-83a9-468b-a56f-db77fe0da375","Type":"ContainerDied","Data":"c7a74ad6a86f92257aa00ff4affbe806c2b2dbe46762f79a5bb253b6ec3388c1"} Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.821422 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a74ad6a86f92257aa00ff4affbe806c2b2dbe46762f79a5bb253b6ec3388c1" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.821485 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.930488 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62"] Nov 23 20:49:59 crc kubenswrapper[4726]: E1123 20:49:59.930840 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e5cdd7-83a9-468b-a56f-db77fe0da375" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.930879 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e5cdd7-83a9-468b-a56f-db77fe0da375" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.931129 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e5cdd7-83a9-468b-a56f-db77fe0da375" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.931695 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.935190 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.935481 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.935834 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.936233 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.937543 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.940710 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:49:59 crc kubenswrapper[4726]: I1123 20:49:59.945103 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62"] Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.025180 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.025218 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4t52\" (UniqueName: \"kubernetes.io/projected/09d671dd-9524-4d36-8db8-a178093fdc70-kube-api-access-p4t52\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.025263 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.025289 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/09d671dd-9524-4d36-8db8-a178093fdc70-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.025358 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.025412 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.126393 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.126539 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.126567 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4t52\" (UniqueName: \"kubernetes.io/projected/09d671dd-9524-4d36-8db8-a178093fdc70-kube-api-access-p4t52\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.126615 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.126652 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/09d671dd-9524-4d36-8db8-a178093fdc70-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.126725 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.127773 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/09d671dd-9524-4d36-8db8-a178093fdc70-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.130816 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.131440 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.133524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.139225 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.147001 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4t52\" (UniqueName: \"kubernetes.io/projected/09d671dd-9524-4d36-8db8-a178093fdc70-kube-api-access-p4t52\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-q6k62\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.265861 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.791390 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62"] Nov 23 20:50:00 crc kubenswrapper[4726]: I1123 20:50:00.834374 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" event={"ID":"09d671dd-9524-4d36-8db8-a178093fdc70","Type":"ContainerStarted","Data":"b40a433d2684dd4cab9bc2de47eea07478fad57c2ebf752759d3c31fcfc40b0e"} Nov 23 20:50:01 crc kubenswrapper[4726]: I1123 20:50:01.846674 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" event={"ID":"09d671dd-9524-4d36-8db8-a178093fdc70","Type":"ContainerStarted","Data":"172cc087a610d42a3d374ac5548fdf1b4fdcab6782ecbad725dc71a444c4cb67"} Nov 23 20:50:01 crc kubenswrapper[4726]: I1123 20:50:01.885200 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" podStartSLOduration=2.451911385 podStartE2EDuration="2.885170315s" podCreationTimestamp="2025-11-23 20:49:59 +0000 UTC" firstStartedPulling="2025-11-23 20:50:00.795313669 +0000 UTC m=+2508.944354655" lastFinishedPulling="2025-11-23 20:50:01.228572589 +0000 UTC m=+2509.377613585" observedRunningTime="2025-11-23 20:50:01.879293329 +0000 UTC m=+2510.028334315" watchObservedRunningTime="2025-11-23 20:50:01.885170315 +0000 UTC m=+2510.034211311" Nov 23 20:51:32 crc kubenswrapper[4726]: I1123 20:51:32.702825 4726 generic.go:334] "Generic (PLEG): container finished" podID="09d671dd-9524-4d36-8db8-a178093fdc70" containerID="172cc087a610d42a3d374ac5548fdf1b4fdcab6782ecbad725dc71a444c4cb67" exitCode=0 Nov 23 20:51:32 crc kubenswrapper[4726]: I1123 20:51:32.702918 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" event={"ID":"09d671dd-9524-4d36-8db8-a178093fdc70","Type":"ContainerDied","Data":"172cc087a610d42a3d374ac5548fdf1b4fdcab6782ecbad725dc71a444c4cb67"} Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.187123 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.381515 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/09d671dd-9524-4d36-8db8-a178093fdc70-ovncontroller-config-0\") pod \"09d671dd-9524-4d36-8db8-a178093fdc70\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.381610 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ssh-key\") pod \"09d671dd-9524-4d36-8db8-a178093fdc70\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.381638 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ovn-combined-ca-bundle\") pod \"09d671dd-9524-4d36-8db8-a178093fdc70\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.381680 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4t52\" (UniqueName: \"kubernetes.io/projected/09d671dd-9524-4d36-8db8-a178093fdc70-kube-api-access-p4t52\") pod \"09d671dd-9524-4d36-8db8-a178093fdc70\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.381723 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-inventory\") pod \"09d671dd-9524-4d36-8db8-a178093fdc70\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.381747 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ceph\") pod \"09d671dd-9524-4d36-8db8-a178093fdc70\" (UID: \"09d671dd-9524-4d36-8db8-a178093fdc70\") " Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.390065 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "09d671dd-9524-4d36-8db8-a178093fdc70" (UID: "09d671dd-9524-4d36-8db8-a178093fdc70"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.390097 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ceph" (OuterVolumeSpecName: "ceph") pod "09d671dd-9524-4d36-8db8-a178093fdc70" (UID: "09d671dd-9524-4d36-8db8-a178093fdc70"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.394999 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09d671dd-9524-4d36-8db8-a178093fdc70-kube-api-access-p4t52" (OuterVolumeSpecName: "kube-api-access-p4t52") pod "09d671dd-9524-4d36-8db8-a178093fdc70" (UID: "09d671dd-9524-4d36-8db8-a178093fdc70"). InnerVolumeSpecName "kube-api-access-p4t52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.408514 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09d671dd-9524-4d36-8db8-a178093fdc70-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "09d671dd-9524-4d36-8db8-a178093fdc70" (UID: "09d671dd-9524-4d36-8db8-a178093fdc70"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.408598 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-inventory" (OuterVolumeSpecName: "inventory") pod "09d671dd-9524-4d36-8db8-a178093fdc70" (UID: "09d671dd-9524-4d36-8db8-a178093fdc70"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.411817 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "09d671dd-9524-4d36-8db8-a178093fdc70" (UID: "09d671dd-9524-4d36-8db8-a178093fdc70"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.484768 4726 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/09d671dd-9524-4d36-8db8-a178093fdc70-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.484821 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.484842 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.484860 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4t52\" (UniqueName: \"kubernetes.io/projected/09d671dd-9524-4d36-8db8-a178093fdc70-kube-api-access-p4t52\") on node \"crc\" DevicePath \"\"" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.484901 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.484918 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/09d671dd-9524-4d36-8db8-a178093fdc70-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.719545 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" event={"ID":"09d671dd-9524-4d36-8db8-a178093fdc70","Type":"ContainerDied","Data":"b40a433d2684dd4cab9bc2de47eea07478fad57c2ebf752759d3c31fcfc40b0e"} Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.719608 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b40a433d2684dd4cab9bc2de47eea07478fad57c2ebf752759d3c31fcfc40b0e" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.719688 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-q6k62" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.843688 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4"] Nov 23 20:51:34 crc kubenswrapper[4726]: E1123 20:51:34.844026 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09d671dd-9524-4d36-8db8-a178093fdc70" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.844042 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="09d671dd-9524-4d36-8db8-a178093fdc70" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.844224 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="09d671dd-9524-4d36-8db8-a178093fdc70" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.844787 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.849338 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.849747 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.850465 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.850798 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.851165 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.851483 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.857742 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.871259 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4"] Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.993166 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.993227 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.993292 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.993322 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x54dz\" (UniqueName: \"kubernetes.io/projected/d233adb5-8473-4545-b91a-56ec95775b67-kube-api-access-x54dz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.993420 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.993507 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:34 crc kubenswrapper[4726]: I1123 20:51:34.993580 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.095001 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.095057 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.095093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x54dz\" (UniqueName: \"kubernetes.io/projected/d233adb5-8473-4545-b91a-56ec95775b67-kube-api-access-x54dz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.095171 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.095240 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.095296 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.095345 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.102614 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.103546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.106467 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.109943 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.113110 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.117902 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.119160 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x54dz\" (UniqueName: \"kubernetes.io/projected/d233adb5-8473-4545-b91a-56ec95775b67-kube-api-access-x54dz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.170525 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:51:35 crc kubenswrapper[4726]: I1123 20:51:35.806182 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4"] Nov 23 20:51:36 crc kubenswrapper[4726]: I1123 20:51:36.737212 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" event={"ID":"d233adb5-8473-4545-b91a-56ec95775b67","Type":"ContainerStarted","Data":"94fbe678c1bde4af281f1cd4d9e4d3b1c6ecbe2c9d42b996bc21eebbd37004ac"} Nov 23 20:51:37 crc kubenswrapper[4726]: I1123 20:51:37.756999 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" event={"ID":"d233adb5-8473-4545-b91a-56ec95775b67","Type":"ContainerStarted","Data":"2d9b512465b41899cac10b0e2e2c71a2498c174e95a44fb8cff64948ac6ad76d"} Nov 23 20:51:37 crc kubenswrapper[4726]: I1123 20:51:37.790487 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" podStartSLOduration=2.862804133 podStartE2EDuration="3.790462368s" podCreationTimestamp="2025-11-23 20:51:34 +0000 UTC" firstStartedPulling="2025-11-23 20:51:35.810733898 +0000 UTC m=+2603.959774854" lastFinishedPulling="2025-11-23 20:51:36.738392133 +0000 UTC m=+2604.887433089" observedRunningTime="2025-11-23 20:51:37.77165588 +0000 UTC m=+2605.920696846" watchObservedRunningTime="2025-11-23 20:51:37.790462368 +0000 UTC m=+2605.939503314" Nov 23 20:52:09 crc kubenswrapper[4726]: I1123 20:52:09.043466 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:52:09 crc kubenswrapper[4726]: I1123 20:52:09.044020 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:52:39 crc kubenswrapper[4726]: I1123 20:52:39.042794 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:52:39 crc kubenswrapper[4726]: I1123 20:52:39.044021 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:52:51 crc kubenswrapper[4726]: I1123 20:52:51.456412 4726 generic.go:334] "Generic (PLEG): container finished" podID="d233adb5-8473-4545-b91a-56ec95775b67" containerID="2d9b512465b41899cac10b0e2e2c71a2498c174e95a44fb8cff64948ac6ad76d" exitCode=0 Nov 23 20:52:51 crc kubenswrapper[4726]: I1123 20:52:51.456467 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" event={"ID":"d233adb5-8473-4545-b91a-56ec95775b67","Type":"ContainerDied","Data":"2d9b512465b41899cac10b0e2e2c71a2498c174e95a44fb8cff64948ac6ad76d"} Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.801673 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.909808 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ceph\") pod \"d233adb5-8473-4545-b91a-56ec95775b67\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.910269 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-inventory\") pod \"d233adb5-8473-4545-b91a-56ec95775b67\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.910435 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-metadata-combined-ca-bundle\") pod \"d233adb5-8473-4545-b91a-56ec95775b67\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.910462 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-nova-metadata-neutron-config-0\") pod \"d233adb5-8473-4545-b91a-56ec95775b67\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.910995 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-ovn-metadata-agent-neutron-config-0\") pod \"d233adb5-8473-4545-b91a-56ec95775b67\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.911049 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x54dz\" (UniqueName: \"kubernetes.io/projected/d233adb5-8473-4545-b91a-56ec95775b67-kube-api-access-x54dz\") pod \"d233adb5-8473-4545-b91a-56ec95775b67\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.911075 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ssh-key\") pod \"d233adb5-8473-4545-b91a-56ec95775b67\" (UID: \"d233adb5-8473-4545-b91a-56ec95775b67\") " Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.917029 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ceph" (OuterVolumeSpecName: "ceph") pod "d233adb5-8473-4545-b91a-56ec95775b67" (UID: "d233adb5-8473-4545-b91a-56ec95775b67"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.917050 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d233adb5-8473-4545-b91a-56ec95775b67" (UID: "d233adb5-8473-4545-b91a-56ec95775b67"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.923262 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d233adb5-8473-4545-b91a-56ec95775b67-kube-api-access-x54dz" (OuterVolumeSpecName: "kube-api-access-x54dz") pod "d233adb5-8473-4545-b91a-56ec95775b67" (UID: "d233adb5-8473-4545-b91a-56ec95775b67"). InnerVolumeSpecName "kube-api-access-x54dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.936062 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "d233adb5-8473-4545-b91a-56ec95775b67" (UID: "d233adb5-8473-4545-b91a-56ec95775b67"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.938854 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-inventory" (OuterVolumeSpecName: "inventory") pod "d233adb5-8473-4545-b91a-56ec95775b67" (UID: "d233adb5-8473-4545-b91a-56ec95775b67"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.939248 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "d233adb5-8473-4545-b91a-56ec95775b67" (UID: "d233adb5-8473-4545-b91a-56ec95775b67"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:52:52 crc kubenswrapper[4726]: I1123 20:52:52.945253 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d233adb5-8473-4545-b91a-56ec95775b67" (UID: "d233adb5-8473-4545-b91a-56ec95775b67"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.014047 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.014087 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x54dz\" (UniqueName: \"kubernetes.io/projected/d233adb5-8473-4545-b91a-56ec95775b67-kube-api-access-x54dz\") on node \"crc\" DevicePath \"\"" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.014100 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.014112 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.014124 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.014135 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.014149 4726 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d233adb5-8473-4545-b91a-56ec95775b67-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.474719 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" event={"ID":"d233adb5-8473-4545-b91a-56ec95775b67","Type":"ContainerDied","Data":"94fbe678c1bde4af281f1cd4d9e4d3b1c6ecbe2c9d42b996bc21eebbd37004ac"} Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.474759 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94fbe678c1bde4af281f1cd4d9e4d3b1c6ecbe2c9d42b996bc21eebbd37004ac" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.474775 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.594990 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n"] Nov 23 20:52:53 crc kubenswrapper[4726]: E1123 20:52:53.598997 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d233adb5-8473-4545-b91a-56ec95775b67" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.599036 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d233adb5-8473-4545-b91a-56ec95775b67" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.599302 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d233adb5-8473-4545-b91a-56ec95775b67" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.600101 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.606595 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.606703 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.607017 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.607227 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.608453 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.608636 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.622368 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n"] Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.724113 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.724173 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.724198 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmzs\" (UniqueName: \"kubernetes.io/projected/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-kube-api-access-8cmzs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.724486 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.724638 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.724677 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.826783 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.826881 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.826907 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.826947 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.826985 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.827004 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmzs\" (UniqueName: \"kubernetes.io/projected/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-kube-api-access-8cmzs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.844829 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.849349 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.856454 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.856693 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.856820 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.857417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmzs\" (UniqueName: \"kubernetes.io/projected/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-kube-api-access-8cmzs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:53 crc kubenswrapper[4726]: I1123 20:52:53.916164 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:52:54 crc kubenswrapper[4726]: I1123 20:52:54.467624 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n"] Nov 23 20:52:55 crc kubenswrapper[4726]: I1123 20:52:55.499449 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" event={"ID":"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91","Type":"ContainerStarted","Data":"d8cad5ee1371f4c257f0735f09ed2f0958b4f7417af5c1692dd9512024a6c58e"} Nov 23 20:52:55 crc kubenswrapper[4726]: I1123 20:52:55.499723 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" event={"ID":"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91","Type":"ContainerStarted","Data":"649edcf2f1858688d647e99b11fc02989d43d8cce50933656d3b63629a361bde"} Nov 23 20:52:55 crc kubenswrapper[4726]: I1123 20:52:55.525862 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" podStartSLOduration=2.09977747 podStartE2EDuration="2.525845533s" podCreationTimestamp="2025-11-23 20:52:53 +0000 UTC" firstStartedPulling="2025-11-23 20:52:54.487787782 +0000 UTC m=+2682.636828738" lastFinishedPulling="2025-11-23 20:52:54.913855845 +0000 UTC m=+2683.062896801" observedRunningTime="2025-11-23 20:52:55.522660344 +0000 UTC m=+2683.671701370" watchObservedRunningTime="2025-11-23 20:52:55.525845533 +0000 UTC m=+2683.674886489" Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.043146 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.043558 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.043632 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.044270 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f7ac676ac424776f668164058671554206e267544abf2ca6f34ada9645d7798"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.044325 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://5f7ac676ac424776f668164058671554206e267544abf2ca6f34ada9645d7798" gracePeriod=600 Nov 23 20:53:09 crc kubenswrapper[4726]: E1123 20:53:09.231899 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e3ac186_9f76_4774_8e04_fb00add1eb72.slice/crio-conmon-5f7ac676ac424776f668164058671554206e267544abf2ca6f34ada9645d7798.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e3ac186_9f76_4774_8e04_fb00add1eb72.slice/crio-5f7ac676ac424776f668164058671554206e267544abf2ca6f34ada9645d7798.scope\": RecentStats: unable to find data in memory cache]" Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.636807 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="5f7ac676ac424776f668164058671554206e267544abf2ca6f34ada9645d7798" exitCode=0 Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.636858 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"5f7ac676ac424776f668164058671554206e267544abf2ca6f34ada9645d7798"} Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.636966 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4"} Nov 23 20:53:09 crc kubenswrapper[4726]: I1123 20:53:09.637019 4726 scope.go:117] "RemoveContainer" containerID="95e660f80e6cafd7cb0492c21c78fdda5c673fbeaf51ff49779ef4aba9126d5e" Nov 23 20:53:20 crc kubenswrapper[4726]: I1123 20:53:20.910486 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h6jxm"] Nov 23 20:53:20 crc kubenswrapper[4726]: I1123 20:53:20.913258 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:20 crc kubenswrapper[4726]: I1123 20:53:20.939209 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6jxm"] Nov 23 20:53:20 crc kubenswrapper[4726]: I1123 20:53:20.969563 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-catalog-content\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:20 crc kubenswrapper[4726]: I1123 20:53:20.969610 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6thd\" (UniqueName: \"kubernetes.io/projected/5db16cf1-9a62-4e6e-87c1-22324265ff8d-kube-api-access-x6thd\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:20 crc kubenswrapper[4726]: I1123 20:53:20.969682 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-utilities\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:21 crc kubenswrapper[4726]: I1123 20:53:21.071810 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-utilities\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:21 crc kubenswrapper[4726]: I1123 20:53:21.072021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-catalog-content\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:21 crc kubenswrapper[4726]: I1123 20:53:21.072051 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6thd\" (UniqueName: \"kubernetes.io/projected/5db16cf1-9a62-4e6e-87c1-22324265ff8d-kube-api-access-x6thd\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:21 crc kubenswrapper[4726]: I1123 20:53:21.072491 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-utilities\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:21 crc kubenswrapper[4726]: I1123 20:53:21.072611 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-catalog-content\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:21 crc kubenswrapper[4726]: I1123 20:53:21.097002 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6thd\" (UniqueName: \"kubernetes.io/projected/5db16cf1-9a62-4e6e-87c1-22324265ff8d-kube-api-access-x6thd\") pod \"redhat-marketplace-h6jxm\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:21 crc kubenswrapper[4726]: I1123 20:53:21.291975 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:21 crc kubenswrapper[4726]: I1123 20:53:21.758330 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6jxm"] Nov 23 20:53:22 crc kubenswrapper[4726]: I1123 20:53:22.737737 4726 generic.go:334] "Generic (PLEG): container finished" podID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerID="2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229" exitCode=0 Nov 23 20:53:22 crc kubenswrapper[4726]: I1123 20:53:22.737780 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6jxm" event={"ID":"5db16cf1-9a62-4e6e-87c1-22324265ff8d","Type":"ContainerDied","Data":"2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229"} Nov 23 20:53:22 crc kubenswrapper[4726]: I1123 20:53:22.737808 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6jxm" event={"ID":"5db16cf1-9a62-4e6e-87c1-22324265ff8d","Type":"ContainerStarted","Data":"9892a204a19f6c2ee35a03fbe4a6240780c044f41b229fbe4f144df840b30b04"} Nov 23 20:53:23 crc kubenswrapper[4726]: I1123 20:53:23.748374 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6jxm" event={"ID":"5db16cf1-9a62-4e6e-87c1-22324265ff8d","Type":"ContainerStarted","Data":"dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6"} Nov 23 20:53:24 crc kubenswrapper[4726]: I1123 20:53:24.759531 4726 generic.go:334] "Generic (PLEG): container finished" podID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerID="dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6" exitCode=0 Nov 23 20:53:24 crc kubenswrapper[4726]: I1123 20:53:24.759626 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6jxm" event={"ID":"5db16cf1-9a62-4e6e-87c1-22324265ff8d","Type":"ContainerDied","Data":"dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6"} Nov 23 20:53:25 crc kubenswrapper[4726]: I1123 20:53:25.769058 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6jxm" event={"ID":"5db16cf1-9a62-4e6e-87c1-22324265ff8d","Type":"ContainerStarted","Data":"8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9"} Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.280304 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h6jxm" podStartSLOduration=3.791510023 podStartE2EDuration="6.280282363s" podCreationTimestamp="2025-11-23 20:53:20 +0000 UTC" firstStartedPulling="2025-11-23 20:53:22.739572461 +0000 UTC m=+2710.888613417" lastFinishedPulling="2025-11-23 20:53:25.228344791 +0000 UTC m=+2713.377385757" observedRunningTime="2025-11-23 20:53:25.791682816 +0000 UTC m=+2713.940723772" watchObservedRunningTime="2025-11-23 20:53:26.280282363 +0000 UTC m=+2714.429323329" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.288479 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-975ld"] Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.290361 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.300611 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-975ld"] Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.399096 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-utilities\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.399155 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-catalog-content\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.399189 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wbv2\" (UniqueName: \"kubernetes.io/projected/5681afcc-0685-4bfb-95cb-a5c086834b2f-kube-api-access-4wbv2\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.502132 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-utilities\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.502186 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-catalog-content\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.502219 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wbv2\" (UniqueName: \"kubernetes.io/projected/5681afcc-0685-4bfb-95cb-a5c086834b2f-kube-api-access-4wbv2\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.502669 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-utilities\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.502725 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-catalog-content\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.523339 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wbv2\" (UniqueName: \"kubernetes.io/projected/5681afcc-0685-4bfb-95cb-a5c086834b2f-kube-api-access-4wbv2\") pod \"redhat-operators-975ld\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:26 crc kubenswrapper[4726]: I1123 20:53:26.617521 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:27 crc kubenswrapper[4726]: I1123 20:53:27.106746 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-975ld"] Nov 23 20:53:27 crc kubenswrapper[4726]: I1123 20:53:27.784315 4726 generic.go:334] "Generic (PLEG): container finished" podID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerID="06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f" exitCode=0 Nov 23 20:53:27 crc kubenswrapper[4726]: I1123 20:53:27.784365 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-975ld" event={"ID":"5681afcc-0685-4bfb-95cb-a5c086834b2f","Type":"ContainerDied","Data":"06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f"} Nov 23 20:53:27 crc kubenswrapper[4726]: I1123 20:53:27.784660 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-975ld" event={"ID":"5681afcc-0685-4bfb-95cb-a5c086834b2f","Type":"ContainerStarted","Data":"c435b40aacf44ff33af6ba14fa9336494f1465971a95c0d6b569188b5f52dc58"} Nov 23 20:53:27 crc kubenswrapper[4726]: I1123 20:53:27.785906 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 20:53:28 crc kubenswrapper[4726]: I1123 20:53:28.796699 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-975ld" event={"ID":"5681afcc-0685-4bfb-95cb-a5c086834b2f","Type":"ContainerStarted","Data":"64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638"} Nov 23 20:53:31 crc kubenswrapper[4726]: I1123 20:53:31.292564 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:31 crc kubenswrapper[4726]: I1123 20:53:31.292945 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:31 crc kubenswrapper[4726]: I1123 20:53:31.358785 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:31 crc kubenswrapper[4726]: I1123 20:53:31.876276 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:32 crc kubenswrapper[4726]: I1123 20:53:32.478400 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6jxm"] Nov 23 20:53:32 crc kubenswrapper[4726]: I1123 20:53:32.832123 4726 generic.go:334] "Generic (PLEG): container finished" podID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerID="64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638" exitCode=0 Nov 23 20:53:32 crc kubenswrapper[4726]: I1123 20:53:32.832205 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-975ld" event={"ID":"5681afcc-0685-4bfb-95cb-a5c086834b2f","Type":"ContainerDied","Data":"64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638"} Nov 23 20:53:33 crc kubenswrapper[4726]: I1123 20:53:33.842323 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-975ld" event={"ID":"5681afcc-0685-4bfb-95cb-a5c086834b2f","Type":"ContainerStarted","Data":"7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb"} Nov 23 20:53:33 crc kubenswrapper[4726]: I1123 20:53:33.842468 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h6jxm" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerName="registry-server" containerID="cri-o://8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9" gracePeriod=2 Nov 23 20:53:33 crc kubenswrapper[4726]: I1123 20:53:33.884239 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-975ld" podStartSLOduration=2.443748258 podStartE2EDuration="7.884220057s" podCreationTimestamp="2025-11-23 20:53:26 +0000 UTC" firstStartedPulling="2025-11-23 20:53:27.785682386 +0000 UTC m=+2715.934723342" lastFinishedPulling="2025-11-23 20:53:33.226154185 +0000 UTC m=+2721.375195141" observedRunningTime="2025-11-23 20:53:33.876159661 +0000 UTC m=+2722.025200667" watchObservedRunningTime="2025-11-23 20:53:33.884220057 +0000 UTC m=+2722.033261023" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.317730 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.461216 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-catalog-content\") pod \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.461639 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6thd\" (UniqueName: \"kubernetes.io/projected/5db16cf1-9a62-4e6e-87c1-22324265ff8d-kube-api-access-x6thd\") pod \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.461715 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-utilities\") pod \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\" (UID: \"5db16cf1-9a62-4e6e-87c1-22324265ff8d\") " Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.462779 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-utilities" (OuterVolumeSpecName: "utilities") pod "5db16cf1-9a62-4e6e-87c1-22324265ff8d" (UID: "5db16cf1-9a62-4e6e-87c1-22324265ff8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.473097 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5db16cf1-9a62-4e6e-87c1-22324265ff8d-kube-api-access-x6thd" (OuterVolumeSpecName: "kube-api-access-x6thd") pod "5db16cf1-9a62-4e6e-87c1-22324265ff8d" (UID: "5db16cf1-9a62-4e6e-87c1-22324265ff8d"). InnerVolumeSpecName "kube-api-access-x6thd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.481317 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5db16cf1-9a62-4e6e-87c1-22324265ff8d" (UID: "5db16cf1-9a62-4e6e-87c1-22324265ff8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.565086 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.565151 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5db16cf1-9a62-4e6e-87c1-22324265ff8d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.565178 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6thd\" (UniqueName: \"kubernetes.io/projected/5db16cf1-9a62-4e6e-87c1-22324265ff8d-kube-api-access-x6thd\") on node \"crc\" DevicePath \"\"" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.853716 4726 generic.go:334] "Generic (PLEG): container finished" podID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerID="8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9" exitCode=0 Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.853767 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6jxm" event={"ID":"5db16cf1-9a62-4e6e-87c1-22324265ff8d","Type":"ContainerDied","Data":"8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9"} Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.853798 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6jxm" event={"ID":"5db16cf1-9a62-4e6e-87c1-22324265ff8d","Type":"ContainerDied","Data":"9892a204a19f6c2ee35a03fbe4a6240780c044f41b229fbe4f144df840b30b04"} Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.853819 4726 scope.go:117] "RemoveContainer" containerID="8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.853958 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6jxm" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.881497 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6jxm"] Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.893697 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6jxm"] Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.895621 4726 scope.go:117] "RemoveContainer" containerID="dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.912941 4726 scope.go:117] "RemoveContainer" containerID="2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.965960 4726 scope.go:117] "RemoveContainer" containerID="8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9" Nov 23 20:53:34 crc kubenswrapper[4726]: E1123 20:53:34.966492 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9\": container with ID starting with 8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9 not found: ID does not exist" containerID="8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.966522 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9"} err="failed to get container status \"8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9\": rpc error: code = NotFound desc = could not find container \"8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9\": container with ID starting with 8253d796e55d9de8135c0fe1e942e061e9d2f56629ac2c5f18398bec1c0a45f9 not found: ID does not exist" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.966544 4726 scope.go:117] "RemoveContainer" containerID="dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6" Nov 23 20:53:34 crc kubenswrapper[4726]: E1123 20:53:34.967062 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6\": container with ID starting with dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6 not found: ID does not exist" containerID="dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.967103 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6"} err="failed to get container status \"dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6\": rpc error: code = NotFound desc = could not find container \"dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6\": container with ID starting with dec9e0f6bf4549939f1caa0d69779b3993ac0f228ed2c55ad9823ce9202d31d6 not found: ID does not exist" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.967133 4726 scope.go:117] "RemoveContainer" containerID="2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229" Nov 23 20:53:34 crc kubenswrapper[4726]: E1123 20:53:34.967587 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229\": container with ID starting with 2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229 not found: ID does not exist" containerID="2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229" Nov 23 20:53:34 crc kubenswrapper[4726]: I1123 20:53:34.967636 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229"} err="failed to get container status \"2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229\": rpc error: code = NotFound desc = could not find container \"2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229\": container with ID starting with 2baf56b65d1b29945f60ead26f732e8c3b0f8375c4dbc28ae036da8a9b4e2229 not found: ID does not exist" Nov 23 20:53:36 crc kubenswrapper[4726]: I1123 20:53:36.602093 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" path="/var/lib/kubelet/pods/5db16cf1-9a62-4e6e-87c1-22324265ff8d/volumes" Nov 23 20:53:36 crc kubenswrapper[4726]: I1123 20:53:36.618197 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:36 crc kubenswrapper[4726]: I1123 20:53:36.618253 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:37 crc kubenswrapper[4726]: I1123 20:53:37.662598 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-975ld" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="registry-server" probeResult="failure" output=< Nov 23 20:53:37 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 20:53:37 crc kubenswrapper[4726]: > Nov 23 20:53:47 crc kubenswrapper[4726]: I1123 20:53:47.672778 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-975ld" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="registry-server" probeResult="failure" output=< Nov 23 20:53:47 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 20:53:47 crc kubenswrapper[4726]: > Nov 23 20:53:56 crc kubenswrapper[4726]: I1123 20:53:56.674320 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:56 crc kubenswrapper[4726]: I1123 20:53:56.724811 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:57 crc kubenswrapper[4726]: I1123 20:53:57.491442 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-975ld"] Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.122975 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-975ld" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="registry-server" containerID="cri-o://7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb" gracePeriod=2 Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.552973 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.702811 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-catalog-content\") pod \"5681afcc-0685-4bfb-95cb-a5c086834b2f\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.703051 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-utilities\") pod \"5681afcc-0685-4bfb-95cb-a5c086834b2f\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.703305 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wbv2\" (UniqueName: \"kubernetes.io/projected/5681afcc-0685-4bfb-95cb-a5c086834b2f-kube-api-access-4wbv2\") pod \"5681afcc-0685-4bfb-95cb-a5c086834b2f\" (UID: \"5681afcc-0685-4bfb-95cb-a5c086834b2f\") " Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.705327 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-utilities" (OuterVolumeSpecName: "utilities") pod "5681afcc-0685-4bfb-95cb-a5c086834b2f" (UID: "5681afcc-0685-4bfb-95cb-a5c086834b2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.710828 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5681afcc-0685-4bfb-95cb-a5c086834b2f-kube-api-access-4wbv2" (OuterVolumeSpecName: "kube-api-access-4wbv2") pod "5681afcc-0685-4bfb-95cb-a5c086834b2f" (UID: "5681afcc-0685-4bfb-95cb-a5c086834b2f"). InnerVolumeSpecName "kube-api-access-4wbv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.807189 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wbv2\" (UniqueName: \"kubernetes.io/projected/5681afcc-0685-4bfb-95cb-a5c086834b2f-kube-api-access-4wbv2\") on node \"crc\" DevicePath \"\"" Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.807222 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.811243 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5681afcc-0685-4bfb-95cb-a5c086834b2f" (UID: "5681afcc-0685-4bfb-95cb-a5c086834b2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:53:58 crc kubenswrapper[4726]: I1123 20:53:58.909421 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5681afcc-0685-4bfb-95cb-a5c086834b2f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.135065 4726 generic.go:334] "Generic (PLEG): container finished" podID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerID="7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb" exitCode=0 Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.135109 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-975ld" event={"ID":"5681afcc-0685-4bfb-95cb-a5c086834b2f","Type":"ContainerDied","Data":"7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb"} Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.135127 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-975ld" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.135141 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-975ld" event={"ID":"5681afcc-0685-4bfb-95cb-a5c086834b2f","Type":"ContainerDied","Data":"c435b40aacf44ff33af6ba14fa9336494f1465971a95c0d6b569188b5f52dc58"} Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.135162 4726 scope.go:117] "RemoveContainer" containerID="7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.163299 4726 scope.go:117] "RemoveContainer" containerID="64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.171614 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-975ld"] Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.190313 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-975ld"] Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.201046 4726 scope.go:117] "RemoveContainer" containerID="06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.235798 4726 scope.go:117] "RemoveContainer" containerID="7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb" Nov 23 20:53:59 crc kubenswrapper[4726]: E1123 20:53:59.236435 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb\": container with ID starting with 7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb not found: ID does not exist" containerID="7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.236487 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb"} err="failed to get container status \"7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb\": rpc error: code = NotFound desc = could not find container \"7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb\": container with ID starting with 7aa6bd2da3893175ad8864bab2fb97250ee4eb02ca4f2cc834ce2476754d16bb not found: ID does not exist" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.236520 4726 scope.go:117] "RemoveContainer" containerID="64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638" Nov 23 20:53:59 crc kubenswrapper[4726]: E1123 20:53:59.237023 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638\": container with ID starting with 64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638 not found: ID does not exist" containerID="64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.237089 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638"} err="failed to get container status \"64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638\": rpc error: code = NotFound desc = could not find container \"64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638\": container with ID starting with 64bf19370213daa258f117ea9ba3a61cbbdee83f8daac5e5cb696e87c6e83638 not found: ID does not exist" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.237131 4726 scope.go:117] "RemoveContainer" containerID="06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f" Nov 23 20:53:59 crc kubenswrapper[4726]: E1123 20:53:59.237575 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f\": container with ID starting with 06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f not found: ID does not exist" containerID="06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f" Nov 23 20:53:59 crc kubenswrapper[4726]: I1123 20:53:59.237606 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f"} err="failed to get container status \"06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f\": rpc error: code = NotFound desc = could not find container \"06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f\": container with ID starting with 06b00a47b988efab6fc0f293e6591cdc3c0cf69debfd1b5b58ff5901623c445f not found: ID does not exist" Nov 23 20:54:00 crc kubenswrapper[4726]: I1123 20:54:00.600678 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" path="/var/lib/kubelet/pods/5681afcc-0685-4bfb-95cb-a5c086834b2f/volumes" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.080898 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7t5n9"] Nov 23 20:54:08 crc kubenswrapper[4726]: E1123 20:54:08.082305 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="extract-utilities" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.082329 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="extract-utilities" Nov 23 20:54:08 crc kubenswrapper[4726]: E1123 20:54:08.082353 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerName="extract-content" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.082376 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerName="extract-content" Nov 23 20:54:08 crc kubenswrapper[4726]: E1123 20:54:08.082387 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="registry-server" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.082396 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="registry-server" Nov 23 20:54:08 crc kubenswrapper[4726]: E1123 20:54:08.082411 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerName="extract-utilities" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.082421 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerName="extract-utilities" Nov 23 20:54:08 crc kubenswrapper[4726]: E1123 20:54:08.082452 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerName="registry-server" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.082463 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerName="registry-server" Nov 23 20:54:08 crc kubenswrapper[4726]: E1123 20:54:08.082479 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="extract-content" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.082487 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="extract-content" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.082709 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5db16cf1-9a62-4e6e-87c1-22324265ff8d" containerName="registry-server" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.082728 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5681afcc-0685-4bfb-95cb-a5c086834b2f" containerName="registry-server" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.084406 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.110569 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7t5n9"] Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.172036 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zssh9\" (UniqueName: \"kubernetes.io/projected/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-kube-api-access-zssh9\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.172158 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-utilities\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.172325 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-catalog-content\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.273756 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-catalog-content\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.274397 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-catalog-content\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.274715 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zssh9\" (UniqueName: \"kubernetes.io/projected/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-kube-api-access-zssh9\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.275376 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-utilities\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.275785 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-utilities\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.300516 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zssh9\" (UniqueName: \"kubernetes.io/projected/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-kube-api-access-zssh9\") pod \"community-operators-7t5n9\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.550893 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:08 crc kubenswrapper[4726]: I1123 20:54:08.963907 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7t5n9"] Nov 23 20:54:09 crc kubenswrapper[4726]: I1123 20:54:09.249974 4726 generic.go:334] "Generic (PLEG): container finished" podID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerID="b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652" exitCode=0 Nov 23 20:54:09 crc kubenswrapper[4726]: I1123 20:54:09.250016 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t5n9" event={"ID":"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2","Type":"ContainerDied","Data":"b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652"} Nov 23 20:54:09 crc kubenswrapper[4726]: I1123 20:54:09.250042 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t5n9" event={"ID":"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2","Type":"ContainerStarted","Data":"638b16eec74c1e1c8909cb7af785e7f66c5526dd6859332d35b8adc5df2274cd"} Nov 23 20:54:10 crc kubenswrapper[4726]: I1123 20:54:10.294960 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t5n9" event={"ID":"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2","Type":"ContainerStarted","Data":"58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1"} Nov 23 20:54:11 crc kubenswrapper[4726]: I1123 20:54:11.308171 4726 generic.go:334] "Generic (PLEG): container finished" podID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerID="58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1" exitCode=0 Nov 23 20:54:11 crc kubenswrapper[4726]: I1123 20:54:11.308229 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t5n9" event={"ID":"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2","Type":"ContainerDied","Data":"58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1"} Nov 23 20:54:12 crc kubenswrapper[4726]: I1123 20:54:12.320324 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t5n9" event={"ID":"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2","Type":"ContainerStarted","Data":"2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536"} Nov 23 20:54:12 crc kubenswrapper[4726]: I1123 20:54:12.351097 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7t5n9" podStartSLOduration=1.880834552 podStartE2EDuration="4.351081233s" podCreationTimestamp="2025-11-23 20:54:08 +0000 UTC" firstStartedPulling="2025-11-23 20:54:09.251729003 +0000 UTC m=+2757.400769959" lastFinishedPulling="2025-11-23 20:54:11.721975644 +0000 UTC m=+2759.871016640" observedRunningTime="2025-11-23 20:54:12.344998923 +0000 UTC m=+2760.494039909" watchObservedRunningTime="2025-11-23 20:54:12.351081233 +0000 UTC m=+2760.500122199" Nov 23 20:54:18 crc kubenswrapper[4726]: I1123 20:54:18.551542 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:18 crc kubenswrapper[4726]: I1123 20:54:18.553680 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:18 crc kubenswrapper[4726]: I1123 20:54:18.608000 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:19 crc kubenswrapper[4726]: I1123 20:54:19.435925 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:19 crc kubenswrapper[4726]: I1123 20:54:19.476697 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7t5n9"] Nov 23 20:54:21 crc kubenswrapper[4726]: I1123 20:54:21.400156 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7t5n9" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerName="registry-server" containerID="cri-o://2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536" gracePeriod=2 Nov 23 20:54:21 crc kubenswrapper[4726]: I1123 20:54:21.842774 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.008787 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zssh9\" (UniqueName: \"kubernetes.io/projected/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-kube-api-access-zssh9\") pod \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.009002 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-catalog-content\") pod \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.009201 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-utilities\") pod \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\" (UID: \"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2\") " Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.010105 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-utilities" (OuterVolumeSpecName: "utilities") pod "77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" (UID: "77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.019073 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-kube-api-access-zssh9" (OuterVolumeSpecName: "kube-api-access-zssh9") pod "77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" (UID: "77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2"). InnerVolumeSpecName "kube-api-access-zssh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.063510 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" (UID: "77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.111289 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.111330 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.111344 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zssh9\" (UniqueName: \"kubernetes.io/projected/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2-kube-api-access-zssh9\") on node \"crc\" DevicePath \"\"" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.411097 4726 generic.go:334] "Generic (PLEG): container finished" podID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerID="2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536" exitCode=0 Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.411162 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7t5n9" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.411180 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t5n9" event={"ID":"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2","Type":"ContainerDied","Data":"2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536"} Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.411249 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7t5n9" event={"ID":"77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2","Type":"ContainerDied","Data":"638b16eec74c1e1c8909cb7af785e7f66c5526dd6859332d35b8adc5df2274cd"} Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.411282 4726 scope.go:117] "RemoveContainer" containerID="2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.458099 4726 scope.go:117] "RemoveContainer" containerID="58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.463562 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7t5n9"] Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.469701 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7t5n9"] Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.489477 4726 scope.go:117] "RemoveContainer" containerID="b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.525418 4726 scope.go:117] "RemoveContainer" containerID="2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536" Nov 23 20:54:22 crc kubenswrapper[4726]: E1123 20:54:22.527963 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536\": container with ID starting with 2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536 not found: ID does not exist" containerID="2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.528140 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536"} err="failed to get container status \"2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536\": rpc error: code = NotFound desc = could not find container \"2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536\": container with ID starting with 2f301be5b501bcd27a6e3d61076c4d87e2dfac1aa491f9ce61e250768d1f6536 not found: ID does not exist" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.528342 4726 scope.go:117] "RemoveContainer" containerID="58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1" Nov 23 20:54:22 crc kubenswrapper[4726]: E1123 20:54:22.528841 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1\": container with ID starting with 58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1 not found: ID does not exist" containerID="58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.528909 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1"} err="failed to get container status \"58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1\": rpc error: code = NotFound desc = could not find container \"58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1\": container with ID starting with 58a1b47a0d1c585cc6edc1cdd68d6ea6cf1add4b37bd869d1e444e2f7e67abb1 not found: ID does not exist" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.528943 4726 scope.go:117] "RemoveContainer" containerID="b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652" Nov 23 20:54:22 crc kubenswrapper[4726]: E1123 20:54:22.529211 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652\": container with ID starting with b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652 not found: ID does not exist" containerID="b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.529237 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652"} err="failed to get container status \"b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652\": rpc error: code = NotFound desc = could not find container \"b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652\": container with ID starting with b91da0e151cc06ad655adacc9bf6e68e2d28f855f7bc2b0ccf65c4718dfd8652 not found: ID does not exist" Nov 23 20:54:22 crc kubenswrapper[4726]: I1123 20:54:22.609694 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" path="/var/lib/kubelet/pods/77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2/volumes" Nov 23 20:55:09 crc kubenswrapper[4726]: I1123 20:55:09.043669 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:55:09 crc kubenswrapper[4726]: I1123 20:55:09.044367 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:55:39 crc kubenswrapper[4726]: I1123 20:55:39.043353 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:55:39 crc kubenswrapper[4726]: I1123 20:55:39.044849 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.043138 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.043610 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.043645 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.044210 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.044267 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" gracePeriod=600 Nov 23 20:56:09 crc kubenswrapper[4726]: E1123 20:56:09.164717 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.423768 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" exitCode=0 Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.423812 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4"} Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.423843 4726 scope.go:117] "RemoveContainer" containerID="5f7ac676ac424776f668164058671554206e267544abf2ca6f34ada9645d7798" Nov 23 20:56:09 crc kubenswrapper[4726]: I1123 20:56:09.424376 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:56:09 crc kubenswrapper[4726]: E1123 20:56:09.424647 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:56:24 crc kubenswrapper[4726]: I1123 20:56:24.589801 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:56:24 crc kubenswrapper[4726]: E1123 20:56:24.590623 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:56:37 crc kubenswrapper[4726]: I1123 20:56:37.589248 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:56:37 crc kubenswrapper[4726]: E1123 20:56:37.591226 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:56:50 crc kubenswrapper[4726]: I1123 20:56:50.590130 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:56:50 crc kubenswrapper[4726]: E1123 20:56:50.591222 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:57:03 crc kubenswrapper[4726]: I1123 20:57:03.592085 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:57:03 crc kubenswrapper[4726]: E1123 20:57:03.595077 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.500598 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r7rpc"] Nov 23 20:57:11 crc kubenswrapper[4726]: E1123 20:57:11.502917 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerName="registry-server" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.502935 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerName="registry-server" Nov 23 20:57:11 crc kubenswrapper[4726]: E1123 20:57:11.502959 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerName="extract-content" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.502967 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerName="extract-content" Nov 23 20:57:11 crc kubenswrapper[4726]: E1123 20:57:11.502978 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerName="extract-utilities" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.502986 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerName="extract-utilities" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.503217 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f80cde-c7b6-4b7d-ae63-7cd71f0dacb2" containerName="registry-server" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.513258 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r7rpc"] Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.513465 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.596122 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpdwz\" (UniqueName: \"kubernetes.io/projected/fa55d610-d803-4f68-ba88-3d0143b60cba-kube-api-access-dpdwz\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.596407 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-catalog-content\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.596537 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-utilities\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.698162 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpdwz\" (UniqueName: \"kubernetes.io/projected/fa55d610-d803-4f68-ba88-3d0143b60cba-kube-api-access-dpdwz\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.698258 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-catalog-content\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.698305 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-utilities\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.698924 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-utilities\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.698993 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-catalog-content\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.715561 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpdwz\" (UniqueName: \"kubernetes.io/projected/fa55d610-d803-4f68-ba88-3d0143b60cba-kube-api-access-dpdwz\") pod \"certified-operators-r7rpc\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:11 crc kubenswrapper[4726]: I1123 20:57:11.838566 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:12 crc kubenswrapper[4726]: I1123 20:57:12.320120 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r7rpc"] Nov 23 20:57:13 crc kubenswrapper[4726]: I1123 20:57:13.054841 4726 generic.go:334] "Generic (PLEG): container finished" podID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerID="c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008" exitCode=0 Nov 23 20:57:13 crc kubenswrapper[4726]: I1123 20:57:13.054958 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7rpc" event={"ID":"fa55d610-d803-4f68-ba88-3d0143b60cba","Type":"ContainerDied","Data":"c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008"} Nov 23 20:57:13 crc kubenswrapper[4726]: I1123 20:57:13.055972 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7rpc" event={"ID":"fa55d610-d803-4f68-ba88-3d0143b60cba","Type":"ContainerStarted","Data":"e51ace0ee45c7dc4b95e6dd6e5f766a51968b796fe05c34b6510845953cbbf1e"} Nov 23 20:57:14 crc kubenswrapper[4726]: I1123 20:57:14.066418 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7rpc" event={"ID":"fa55d610-d803-4f68-ba88-3d0143b60cba","Type":"ContainerStarted","Data":"1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1"} Nov 23 20:57:15 crc kubenswrapper[4726]: I1123 20:57:15.077005 4726 generic.go:334] "Generic (PLEG): container finished" podID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerID="1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1" exitCode=0 Nov 23 20:57:15 crc kubenswrapper[4726]: I1123 20:57:15.077076 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7rpc" event={"ID":"fa55d610-d803-4f68-ba88-3d0143b60cba","Type":"ContainerDied","Data":"1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1"} Nov 23 20:57:15 crc kubenswrapper[4726]: E1123 20:57:15.137573 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa55d610_d803_4f68_ba88_3d0143b60cba.slice/crio-1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1.scope\": RecentStats: unable to find data in memory cache]" Nov 23 20:57:16 crc kubenswrapper[4726]: I1123 20:57:16.088455 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7rpc" event={"ID":"fa55d610-d803-4f68-ba88-3d0143b60cba","Type":"ContainerStarted","Data":"caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373"} Nov 23 20:57:16 crc kubenswrapper[4726]: I1123 20:57:16.132788 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r7rpc" podStartSLOduration=2.679160793 podStartE2EDuration="5.13276418s" podCreationTimestamp="2025-11-23 20:57:11 +0000 UTC" firstStartedPulling="2025-11-23 20:57:13.056315044 +0000 UTC m=+2941.205355990" lastFinishedPulling="2025-11-23 20:57:15.509918431 +0000 UTC m=+2943.658959377" observedRunningTime="2025-11-23 20:57:16.123369917 +0000 UTC m=+2944.272410873" watchObservedRunningTime="2025-11-23 20:57:16.13276418 +0000 UTC m=+2944.281805136" Nov 23 20:57:16 crc kubenswrapper[4726]: I1123 20:57:16.589422 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:57:16 crc kubenswrapper[4726]: E1123 20:57:16.589916 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:57:21 crc kubenswrapper[4726]: I1123 20:57:21.839586 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:21 crc kubenswrapper[4726]: I1123 20:57:21.841199 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:21 crc kubenswrapper[4726]: I1123 20:57:21.902728 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:22 crc kubenswrapper[4726]: I1123 20:57:22.205675 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:22 crc kubenswrapper[4726]: I1123 20:57:22.263157 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r7rpc"] Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.165901 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r7rpc" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerName="registry-server" containerID="cri-o://caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373" gracePeriod=2 Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.601672 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.691722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpdwz\" (UniqueName: \"kubernetes.io/projected/fa55d610-d803-4f68-ba88-3d0143b60cba-kube-api-access-dpdwz\") pod \"fa55d610-d803-4f68-ba88-3d0143b60cba\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.691907 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-utilities\") pod \"fa55d610-d803-4f68-ba88-3d0143b60cba\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.692013 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-catalog-content\") pod \"fa55d610-d803-4f68-ba88-3d0143b60cba\" (UID: \"fa55d610-d803-4f68-ba88-3d0143b60cba\") " Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.693554 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-utilities" (OuterVolumeSpecName: "utilities") pod "fa55d610-d803-4f68-ba88-3d0143b60cba" (UID: "fa55d610-d803-4f68-ba88-3d0143b60cba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.697786 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa55d610-d803-4f68-ba88-3d0143b60cba-kube-api-access-dpdwz" (OuterVolumeSpecName: "kube-api-access-dpdwz") pod "fa55d610-d803-4f68-ba88-3d0143b60cba" (UID: "fa55d610-d803-4f68-ba88-3d0143b60cba"). InnerVolumeSpecName "kube-api-access-dpdwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.794076 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpdwz\" (UniqueName: \"kubernetes.io/projected/fa55d610-d803-4f68-ba88-3d0143b60cba-kube-api-access-dpdwz\") on node \"crc\" DevicePath \"\"" Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.794105 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.920980 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa55d610-d803-4f68-ba88-3d0143b60cba" (UID: "fa55d610-d803-4f68-ba88-3d0143b60cba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 20:57:24 crc kubenswrapper[4726]: I1123 20:57:24.997488 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa55d610-d803-4f68-ba88-3d0143b60cba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.183262 4726 generic.go:334] "Generic (PLEG): container finished" podID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerID="caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373" exitCode=0 Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.183312 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7rpc" event={"ID":"fa55d610-d803-4f68-ba88-3d0143b60cba","Type":"ContainerDied","Data":"caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373"} Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.183343 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7rpc" event={"ID":"fa55d610-d803-4f68-ba88-3d0143b60cba","Type":"ContainerDied","Data":"e51ace0ee45c7dc4b95e6dd6e5f766a51968b796fe05c34b6510845953cbbf1e"} Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.183363 4726 scope.go:117] "RemoveContainer" containerID="caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.183413 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7rpc" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.226381 4726 scope.go:117] "RemoveContainer" containerID="1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.266439 4726 scope.go:117] "RemoveContainer" containerID="c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.273843 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r7rpc"] Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.283972 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r7rpc"] Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.296041 4726 scope.go:117] "RemoveContainer" containerID="caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373" Nov 23 20:57:25 crc kubenswrapper[4726]: E1123 20:57:25.296404 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373\": container with ID starting with caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373 not found: ID does not exist" containerID="caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.296451 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373"} err="failed to get container status \"caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373\": rpc error: code = NotFound desc = could not find container \"caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373\": container with ID starting with caa62114f480727db76510d33d6d452d11fb593367c5b4c1489b90ad0e452373 not found: ID does not exist" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.296494 4726 scope.go:117] "RemoveContainer" containerID="1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1" Nov 23 20:57:25 crc kubenswrapper[4726]: E1123 20:57:25.296724 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1\": container with ID starting with 1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1 not found: ID does not exist" containerID="1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.296749 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1"} err="failed to get container status \"1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1\": rpc error: code = NotFound desc = could not find container \"1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1\": container with ID starting with 1cf7842c4f71ed42967cb8ad58ac926394a42f6885a4930342ede2bf2581c9e1 not found: ID does not exist" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.296761 4726 scope.go:117] "RemoveContainer" containerID="c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008" Nov 23 20:57:25 crc kubenswrapper[4726]: E1123 20:57:25.296979 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008\": container with ID starting with c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008 not found: ID does not exist" containerID="c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008" Nov 23 20:57:25 crc kubenswrapper[4726]: I1123 20:57:25.296997 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008"} err="failed to get container status \"c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008\": rpc error: code = NotFound desc = could not find container \"c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008\": container with ID starting with c373ac9d18081e1ed088c3314b6e781f4e5210aa9d724bb493d67936fe752008 not found: ID does not exist" Nov 23 20:57:25 crc kubenswrapper[4726]: E1123 20:57:25.438049 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa55d610_d803_4f68_ba88_3d0143b60cba.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa55d610_d803_4f68_ba88_3d0143b60cba.slice/crio-e51ace0ee45c7dc4b95e6dd6e5f766a51968b796fe05c34b6510845953cbbf1e\": RecentStats: unable to find data in memory cache]" Nov 23 20:57:26 crc kubenswrapper[4726]: I1123 20:57:26.600397 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" path="/var/lib/kubelet/pods/fa55d610-d803-4f68-ba88-3d0143b60cba/volumes" Nov 23 20:57:29 crc kubenswrapper[4726]: I1123 20:57:29.588811 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:57:29 crc kubenswrapper[4726]: E1123 20:57:29.589367 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:57:42 crc kubenswrapper[4726]: I1123 20:57:42.593653 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:57:42 crc kubenswrapper[4726]: E1123 20:57:42.595561 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:57:55 crc kubenswrapper[4726]: I1123 20:57:55.589018 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:57:55 crc kubenswrapper[4726]: E1123 20:57:55.589791 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:58:04 crc kubenswrapper[4726]: I1123 20:58:04.608056 4726 generic.go:334] "Generic (PLEG): container finished" podID="4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" containerID="d8cad5ee1371f4c257f0735f09ed2f0958b4f7417af5c1692dd9512024a6c58e" exitCode=0 Nov 23 20:58:04 crc kubenswrapper[4726]: I1123 20:58:04.609227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" event={"ID":"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91","Type":"ContainerDied","Data":"d8cad5ee1371f4c257f0735f09ed2f0958b4f7417af5c1692dd9512024a6c58e"} Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.050482 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.175655 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cmzs\" (UniqueName: \"kubernetes.io/projected/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-kube-api-access-8cmzs\") pod \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.175806 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ssh-key\") pod \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.175931 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-inventory\") pod \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.176049 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-combined-ca-bundle\") pod \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.176196 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ceph\") pod \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.176252 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-secret-0\") pod \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\" (UID: \"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91\") " Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.183464 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-kube-api-access-8cmzs" (OuterVolumeSpecName: "kube-api-access-8cmzs") pod "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" (UID: "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91"). InnerVolumeSpecName "kube-api-access-8cmzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.184237 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" (UID: "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.184853 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ceph" (OuterVolumeSpecName: "ceph") pod "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" (UID: "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.213230 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-inventory" (OuterVolumeSpecName: "inventory") pod "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" (UID: "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.215057 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" (UID: "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.238532 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" (UID: "4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.277888 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.277920 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.277932 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.277941 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.277951 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cmzs\" (UniqueName: \"kubernetes.io/projected/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-kube-api-access-8cmzs\") on node \"crc\" DevicePath \"\"" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.277961 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.631005 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" event={"ID":"4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91","Type":"ContainerDied","Data":"649edcf2f1858688d647e99b11fc02989d43d8cce50933656d3b63629a361bde"} Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.631218 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="649edcf2f1858688d647e99b11fc02989d43d8cce50933656d3b63629a361bde" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.631084 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.775750 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj"] Nov 23 20:58:06 crc kubenswrapper[4726]: E1123 20:58:06.776330 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerName="extract-content" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.776405 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerName="extract-content" Nov 23 20:58:06 crc kubenswrapper[4726]: E1123 20:58:06.776471 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerName="extract-utilities" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.776593 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerName="extract-utilities" Nov 23 20:58:06 crc kubenswrapper[4726]: E1123 20:58:06.776681 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.776737 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 23 20:58:06 crc kubenswrapper[4726]: E1123 20:58:06.776797 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerName="registry-server" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.776854 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerName="registry-server" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.777088 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa55d610-d803-4f68-ba88-3d0143b60cba" containerName="registry-server" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.777155 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.777752 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.783732 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.783846 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.786207 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gvprf" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.787074 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.787195 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.787339 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.787528 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.787718 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.789982 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj"] Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.790350 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887205 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887282 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887309 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s4hx\" (UniqueName: \"kubernetes.io/projected/89cb9922-a339-44bc-b1be-e7ffa127cf02-kube-api-access-7s4hx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887333 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887387 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887461 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887631 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887651 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887744 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887762 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.887796 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989623 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989695 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989754 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989781 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989835 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989861 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989907 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989967 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.989995 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.990016 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s4hx\" (UniqueName: \"kubernetes.io/projected/89cb9922-a339-44bc-b1be-e7ffa127cf02-kube-api-access-7s4hx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.990039 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.991310 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.991468 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.995098 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.995117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.995355 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.995610 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.996367 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.997164 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.997923 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:06 crc kubenswrapper[4726]: I1123 20:58:06.998523 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:07 crc kubenswrapper[4726]: I1123 20:58:07.010365 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s4hx\" (UniqueName: \"kubernetes.io/projected/89cb9922-a339-44bc-b1be-e7ffa127cf02-kube-api-access-7s4hx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:07 crc kubenswrapper[4726]: I1123 20:58:07.092211 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 20:58:07 crc kubenswrapper[4726]: I1123 20:58:07.643755 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj"] Nov 23 20:58:08 crc kubenswrapper[4726]: I1123 20:58:08.589822 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:58:08 crc kubenswrapper[4726]: E1123 20:58:08.590747 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:58:08 crc kubenswrapper[4726]: I1123 20:58:08.652015 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" event={"ID":"89cb9922-a339-44bc-b1be-e7ffa127cf02","Type":"ContainerStarted","Data":"f4df348b44a9d6b6e7db8a48a5508a1e4119a486a54e22b22615fca8d7170c56"} Nov 23 20:58:08 crc kubenswrapper[4726]: I1123 20:58:08.652056 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" event={"ID":"89cb9922-a339-44bc-b1be-e7ffa127cf02","Type":"ContainerStarted","Data":"33503fdffdf74124f0b1625a14515ed6ae6dcac959a54dc9f5588d6d01f232bf"} Nov 23 20:58:08 crc kubenswrapper[4726]: I1123 20:58:08.677496 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" podStartSLOduration=2.15430421 podStartE2EDuration="2.677479632s" podCreationTimestamp="2025-11-23 20:58:06 +0000 UTC" firstStartedPulling="2025-11-23 20:58:07.651838438 +0000 UTC m=+2995.800879394" lastFinishedPulling="2025-11-23 20:58:08.17501385 +0000 UTC m=+2996.324054816" observedRunningTime="2025-11-23 20:58:08.6724507 +0000 UTC m=+2996.821491666" watchObservedRunningTime="2025-11-23 20:58:08.677479632 +0000 UTC m=+2996.826520578" Nov 23 20:58:19 crc kubenswrapper[4726]: I1123 20:58:19.588786 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:58:19 crc kubenswrapper[4726]: E1123 20:58:19.590518 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:58:33 crc kubenswrapper[4726]: I1123 20:58:33.589240 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:58:33 crc kubenswrapper[4726]: E1123 20:58:33.590598 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:58:48 crc kubenswrapper[4726]: I1123 20:58:48.590037 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:58:48 crc kubenswrapper[4726]: E1123 20:58:48.590938 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:59:02 crc kubenswrapper[4726]: I1123 20:59:02.594696 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:59:02 crc kubenswrapper[4726]: E1123 20:59:02.595409 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:59:13 crc kubenswrapper[4726]: I1123 20:59:13.590292 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:59:13 crc kubenswrapper[4726]: E1123 20:59:13.591643 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:59:26 crc kubenswrapper[4726]: I1123 20:59:26.590952 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:59:26 crc kubenswrapper[4726]: E1123 20:59:26.591740 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:59:41 crc kubenswrapper[4726]: I1123 20:59:41.589392 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:59:41 crc kubenswrapper[4726]: E1123 20:59:41.590291 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 20:59:52 crc kubenswrapper[4726]: I1123 20:59:52.599360 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 20:59:52 crc kubenswrapper[4726]: E1123 20:59:52.600571 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.154618 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w"] Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.156673 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.159578 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.159983 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.169523 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w"] Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.326654 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8628c7b9-7f42-447d-8772-1148339b3df0-config-volume\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.326745 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8628c7b9-7f42-447d-8772-1148339b3df0-secret-volume\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.326789 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22bl9\" (UniqueName: \"kubernetes.io/projected/8628c7b9-7f42-447d-8772-1148339b3df0-kube-api-access-22bl9\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.428683 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22bl9\" (UniqueName: \"kubernetes.io/projected/8628c7b9-7f42-447d-8772-1148339b3df0-kube-api-access-22bl9\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.429121 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8628c7b9-7f42-447d-8772-1148339b3df0-config-volume\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.430041 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8628c7b9-7f42-447d-8772-1148339b3df0-secret-volume\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.429964 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8628c7b9-7f42-447d-8772-1148339b3df0-config-volume\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.442423 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8628c7b9-7f42-447d-8772-1148339b3df0-secret-volume\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.446624 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22bl9\" (UniqueName: \"kubernetes.io/projected/8628c7b9-7f42-447d-8772-1148339b3df0-kube-api-access-22bl9\") pod \"collect-profiles-29398860-dm25w\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.480486 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:00 crc kubenswrapper[4726]: I1123 21:00:00.968713 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w"] Nov 23 21:00:00 crc kubenswrapper[4726]: W1123 21:00:00.976120 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8628c7b9_7f42_447d_8772_1148339b3df0.slice/crio-0c75060e91b53af844c87ad3766c2fb4c4ebb6159a2402c646a5e72d96eca9e5 WatchSource:0}: Error finding container 0c75060e91b53af844c87ad3766c2fb4c4ebb6159a2402c646a5e72d96eca9e5: Status 404 returned error can't find the container with id 0c75060e91b53af844c87ad3766c2fb4c4ebb6159a2402c646a5e72d96eca9e5 Nov 23 21:00:01 crc kubenswrapper[4726]: I1123 21:00:01.707441 4726 generic.go:334] "Generic (PLEG): container finished" podID="8628c7b9-7f42-447d-8772-1148339b3df0" containerID="b697200cbe19ab5f59e5c0344910aead9f93083c79bd63a8044c5ecfb886cd44" exitCode=0 Nov 23 21:00:01 crc kubenswrapper[4726]: I1123 21:00:01.707499 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" event={"ID":"8628c7b9-7f42-447d-8772-1148339b3df0","Type":"ContainerDied","Data":"b697200cbe19ab5f59e5c0344910aead9f93083c79bd63a8044c5ecfb886cd44"} Nov 23 21:00:01 crc kubenswrapper[4726]: I1123 21:00:01.707721 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" event={"ID":"8628c7b9-7f42-447d-8772-1148339b3df0","Type":"ContainerStarted","Data":"0c75060e91b53af844c87ad3766c2fb4c4ebb6159a2402c646a5e72d96eca9e5"} Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.058522 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.084603 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8628c7b9-7f42-447d-8772-1148339b3df0-secret-volume\") pod \"8628c7b9-7f42-447d-8772-1148339b3df0\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.084693 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8628c7b9-7f42-447d-8772-1148339b3df0-config-volume\") pod \"8628c7b9-7f42-447d-8772-1148339b3df0\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.084822 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22bl9\" (UniqueName: \"kubernetes.io/projected/8628c7b9-7f42-447d-8772-1148339b3df0-kube-api-access-22bl9\") pod \"8628c7b9-7f42-447d-8772-1148339b3df0\" (UID: \"8628c7b9-7f42-447d-8772-1148339b3df0\") " Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.086166 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8628c7b9-7f42-447d-8772-1148339b3df0-config-volume" (OuterVolumeSpecName: "config-volume") pod "8628c7b9-7f42-447d-8772-1148339b3df0" (UID: "8628c7b9-7f42-447d-8772-1148339b3df0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.122822 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8628c7b9-7f42-447d-8772-1148339b3df0-kube-api-access-22bl9" (OuterVolumeSpecName: "kube-api-access-22bl9") pod "8628c7b9-7f42-447d-8772-1148339b3df0" (UID: "8628c7b9-7f42-447d-8772-1148339b3df0"). InnerVolumeSpecName "kube-api-access-22bl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.128074 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8628c7b9-7f42-447d-8772-1148339b3df0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8628c7b9-7f42-447d-8772-1148339b3df0" (UID: "8628c7b9-7f42-447d-8772-1148339b3df0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.186788 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8628c7b9-7f42-447d-8772-1148339b3df0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.186818 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8628c7b9-7f42-447d-8772-1148339b3df0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.186828 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22bl9\" (UniqueName: \"kubernetes.io/projected/8628c7b9-7f42-447d-8772-1148339b3df0-kube-api-access-22bl9\") on node \"crc\" DevicePath \"\"" Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.727353 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" event={"ID":"8628c7b9-7f42-447d-8772-1148339b3df0","Type":"ContainerDied","Data":"0c75060e91b53af844c87ad3766c2fb4c4ebb6159a2402c646a5e72d96eca9e5"} Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.727645 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c75060e91b53af844c87ad3766c2fb4c4ebb6159a2402c646a5e72d96eca9e5" Nov 23 21:00:03 crc kubenswrapper[4726]: I1123 21:00:03.727473 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398860-dm25w" Nov 23 21:00:04 crc kubenswrapper[4726]: I1123 21:00:04.141709 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj"] Nov 23 21:00:04 crc kubenswrapper[4726]: I1123 21:00:04.153439 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398815-bh5mj"] Nov 23 21:00:04 crc kubenswrapper[4726]: I1123 21:00:04.601022 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cef3ad9-bcdd-417e-85d1-972a8ed016a5" path="/var/lib/kubelet/pods/6cef3ad9-bcdd-417e-85d1-972a8ed016a5/volumes" Nov 23 21:00:07 crc kubenswrapper[4726]: I1123 21:00:07.589447 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 21:00:07 crc kubenswrapper[4726]: E1123 21:00:07.589951 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:00:20 crc kubenswrapper[4726]: I1123 21:00:20.382166 4726 scope.go:117] "RemoveContainer" containerID="04f28e575b48de59825ef33c0d0c6f9a275917c035a138d90859f704bdd06126" Nov 23 21:00:22 crc kubenswrapper[4726]: I1123 21:00:22.598162 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 21:00:22 crc kubenswrapper[4726]: E1123 21:00:22.598810 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:00:35 crc kubenswrapper[4726]: I1123 21:00:35.589608 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 21:00:35 crc kubenswrapper[4726]: E1123 21:00:35.590637 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:00:48 crc kubenswrapper[4726]: I1123 21:00:48.589129 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 21:00:48 crc kubenswrapper[4726]: E1123 21:00:48.589894 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.158251 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29398861-kdlcf"] Nov 23 21:01:00 crc kubenswrapper[4726]: E1123 21:01:00.159762 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8628c7b9-7f42-447d-8772-1148339b3df0" containerName="collect-profiles" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.159783 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8628c7b9-7f42-447d-8772-1148339b3df0" containerName="collect-profiles" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.160165 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8628c7b9-7f42-447d-8772-1148339b3df0" containerName="collect-profiles" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.161533 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.170845 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29398861-kdlcf"] Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.256382 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-config-data\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.256417 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qs72\" (UniqueName: \"kubernetes.io/projected/d2340c37-0f69-48f7-94c1-e417bb7844a5-kube-api-access-4qs72\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.256435 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-fernet-keys\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.256461 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-combined-ca-bundle\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.357969 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-config-data\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.358016 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-fernet-keys\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.358034 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qs72\" (UniqueName: \"kubernetes.io/projected/d2340c37-0f69-48f7-94c1-e417bb7844a5-kube-api-access-4qs72\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.358056 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-combined-ca-bundle\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.367922 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-combined-ca-bundle\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.368281 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-config-data\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.377205 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qs72\" (UniqueName: \"kubernetes.io/projected/d2340c37-0f69-48f7-94c1-e417bb7844a5-kube-api-access-4qs72\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.381119 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-fernet-keys\") pod \"keystone-cron-29398861-kdlcf\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.512432 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:00 crc kubenswrapper[4726]: I1123 21:01:00.955401 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29398861-kdlcf"] Nov 23 21:01:01 crc kubenswrapper[4726]: I1123 21:01:01.287221 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29398861-kdlcf" event={"ID":"d2340c37-0f69-48f7-94c1-e417bb7844a5","Type":"ContainerStarted","Data":"804b796b19b48a3c1b58b505b108e17296e465ddaf1c6da79acdd9f7999f10ea"} Nov 23 21:01:01 crc kubenswrapper[4726]: I1123 21:01:01.287282 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29398861-kdlcf" event={"ID":"d2340c37-0f69-48f7-94c1-e417bb7844a5","Type":"ContainerStarted","Data":"6d0bcdf5bea62bfbd38103db8a9e9e1e5e98b0c6620dc6cc318a1d45b4f3d702"} Nov 23 21:01:01 crc kubenswrapper[4726]: I1123 21:01:01.302509 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29398861-kdlcf" podStartSLOduration=1.302491672 podStartE2EDuration="1.302491672s" podCreationTimestamp="2025-11-23 21:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:01:01.299408856 +0000 UTC m=+3169.448449852" watchObservedRunningTime="2025-11-23 21:01:01.302491672 +0000 UTC m=+3169.451532628" Nov 23 21:01:01 crc kubenswrapper[4726]: I1123 21:01:01.589506 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 21:01:01 crc kubenswrapper[4726]: E1123 21:01:01.590052 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:01:04 crc kubenswrapper[4726]: I1123 21:01:04.319241 4726 generic.go:334] "Generic (PLEG): container finished" podID="d2340c37-0f69-48f7-94c1-e417bb7844a5" containerID="804b796b19b48a3c1b58b505b108e17296e465ddaf1c6da79acdd9f7999f10ea" exitCode=0 Nov 23 21:01:04 crc kubenswrapper[4726]: I1123 21:01:04.319325 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29398861-kdlcf" event={"ID":"d2340c37-0f69-48f7-94c1-e417bb7844a5","Type":"ContainerDied","Data":"804b796b19b48a3c1b58b505b108e17296e465ddaf1c6da79acdd9f7999f10ea"} Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.656583 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.675827 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qs72\" (UniqueName: \"kubernetes.io/projected/d2340c37-0f69-48f7-94c1-e417bb7844a5-kube-api-access-4qs72\") pod \"d2340c37-0f69-48f7-94c1-e417bb7844a5\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.675890 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-fernet-keys\") pod \"d2340c37-0f69-48f7-94c1-e417bb7844a5\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.675958 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-config-data\") pod \"d2340c37-0f69-48f7-94c1-e417bb7844a5\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.675990 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-combined-ca-bundle\") pod \"d2340c37-0f69-48f7-94c1-e417bb7844a5\" (UID: \"d2340c37-0f69-48f7-94c1-e417bb7844a5\") " Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.684542 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2340c37-0f69-48f7-94c1-e417bb7844a5-kube-api-access-4qs72" (OuterVolumeSpecName: "kube-api-access-4qs72") pod "d2340c37-0f69-48f7-94c1-e417bb7844a5" (UID: "d2340c37-0f69-48f7-94c1-e417bb7844a5"). InnerVolumeSpecName "kube-api-access-4qs72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.690827 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d2340c37-0f69-48f7-94c1-e417bb7844a5" (UID: "d2340c37-0f69-48f7-94c1-e417bb7844a5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.716968 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2340c37-0f69-48f7-94c1-e417bb7844a5" (UID: "d2340c37-0f69-48f7-94c1-e417bb7844a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.722811 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-config-data" (OuterVolumeSpecName: "config-data") pod "d2340c37-0f69-48f7-94c1-e417bb7844a5" (UID: "d2340c37-0f69-48f7-94c1-e417bb7844a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.777827 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.777880 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.777900 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qs72\" (UniqueName: \"kubernetes.io/projected/d2340c37-0f69-48f7-94c1-e417bb7844a5-kube-api-access-4qs72\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:05 crc kubenswrapper[4726]: I1123 21:01:05.777913 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2340c37-0f69-48f7-94c1-e417bb7844a5-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:06 crc kubenswrapper[4726]: I1123 21:01:06.340064 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29398861-kdlcf" event={"ID":"d2340c37-0f69-48f7-94c1-e417bb7844a5","Type":"ContainerDied","Data":"6d0bcdf5bea62bfbd38103db8a9e9e1e5e98b0c6620dc6cc318a1d45b4f3d702"} Nov 23 21:01:06 crc kubenswrapper[4726]: I1123 21:01:06.340338 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d0bcdf5bea62bfbd38103db8a9e9e1e5e98b0c6620dc6cc318a1d45b4f3d702" Nov 23 21:01:06 crc kubenswrapper[4726]: I1123 21:01:06.340139 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29398861-kdlcf" Nov 23 21:01:16 crc kubenswrapper[4726]: I1123 21:01:16.590393 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 21:01:17 crc kubenswrapper[4726]: I1123 21:01:17.483433 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"8ce91d02edffb0448f1e3ab904f83011317aa44215088577adc4f62202e8da65"} Nov 23 21:01:57 crc kubenswrapper[4726]: I1123 21:01:57.950964 4726 generic.go:334] "Generic (PLEG): container finished" podID="89cb9922-a339-44bc-b1be-e7ffa127cf02" containerID="f4df348b44a9d6b6e7db8a48a5508a1e4119a486a54e22b22615fca8d7170c56" exitCode=0 Nov 23 21:01:57 crc kubenswrapper[4726]: I1123 21:01:57.951073 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" event={"ID":"89cb9922-a339-44bc-b1be-e7ffa127cf02","Type":"ContainerDied","Data":"f4df348b44a9d6b6e7db8a48a5508a1e4119a486a54e22b22615fca8d7170c56"} Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.388088 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.497686 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph-nova-0\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.497736 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-1\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.497803 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-0\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.497945 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.497976 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s4hx\" (UniqueName: \"kubernetes.io/projected/89cb9922-a339-44bc-b1be-e7ffa127cf02-kube-api-access-7s4hx\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.498002 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-1\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.498025 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-extra-config-0\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.498049 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-inventory\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.498114 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-0\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.498240 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ssh-key\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.498296 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-custom-ceph-combined-ca-bundle\") pod \"89cb9922-a339-44bc-b1be-e7ffa127cf02\" (UID: \"89cb9922-a339-44bc-b1be-e7ffa127cf02\") " Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.507560 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph" (OuterVolumeSpecName: "ceph") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.519981 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.520160 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89cb9922-a339-44bc-b1be-e7ffa127cf02-kube-api-access-7s4hx" (OuterVolumeSpecName: "kube-api-access-7s4hx") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "kube-api-access-7s4hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.528187 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-inventory" (OuterVolumeSpecName: "inventory") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.529940 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.535115 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.538786 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.555032 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.556534 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.560248 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.570770 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "89cb9922-a339-44bc-b1be-e7ffa127cf02" (UID: "89cb9922-a339-44bc-b1be-e7ffa127cf02"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599790 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599827 4726 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599844 4726 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599858 4726 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599884 4726 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599892 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599901 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s4hx\" (UniqueName: \"kubernetes.io/projected/89cb9922-a339-44bc-b1be-e7ffa127cf02-kube-api-access-7s4hx\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599910 4726 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599918 4726 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599926 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-inventory\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.599934 4726 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/89cb9922-a339-44bc-b1be-e7ffa127cf02-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.975955 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" event={"ID":"89cb9922-a339-44bc-b1be-e7ffa127cf02","Type":"ContainerDied","Data":"33503fdffdf74124f0b1625a14515ed6ae6dcac959a54dc9f5588d6d01f232bf"} Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.976372 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33503fdffdf74124f0b1625a14515ed6ae6dcac959a54dc9f5588d6d01f232bf" Nov 23 21:01:59 crc kubenswrapper[4726]: I1123 21:01:59.976024 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.941728 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 23 21:02:14 crc kubenswrapper[4726]: E1123 21:02:14.942686 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2340c37-0f69-48f7-94c1-e417bb7844a5" containerName="keystone-cron" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.942701 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2340c37-0f69-48f7-94c1-e417bb7844a5" containerName="keystone-cron" Nov 23 21:02:14 crc kubenswrapper[4726]: E1123 21:02:14.942746 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cb9922-a339-44bc-b1be-e7ffa127cf02" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.942755 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cb9922-a339-44bc-b1be-e7ffa127cf02" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.944088 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2340c37-0f69-48f7-94c1-e417bb7844a5" containerName="keystone-cron" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.944110 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="89cb9922-a339-44bc-b1be-e7ffa127cf02" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.945099 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.950034 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.950239 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 23 21:02:14 crc kubenswrapper[4726]: I1123 21:02:14.968279 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.019954 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.021516 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.030783 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.031127 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.035840 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-scripts\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.036053 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.036174 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-config-data-custom\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.039547 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.039656 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-ceph\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.039752 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.039831 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.039967 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040151 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-nvme\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040228 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-run\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040312 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040394 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040464 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-run\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040554 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040643 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040736 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040881 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-dev\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.040969 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041051 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041079 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041104 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-config-data\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041136 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041168 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l76s\" (UniqueName: \"kubernetes.io/projected/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-kube-api-access-2l76s\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041193 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-sys\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041256 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041328 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041403 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r29l\" (UniqueName: \"kubernetes.io/projected/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-kube-api-access-9r29l\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041421 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041442 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041472 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-lib-modules\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041497 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.041528 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143610 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143638 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-dev\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143670 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143704 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143720 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143737 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-config-data\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143755 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143769 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l76s\" (UniqueName: \"kubernetes.io/projected/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-kube-api-access-2l76s\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143786 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-sys\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143805 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143824 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143850 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r29l\" (UniqueName: \"kubernetes.io/projected/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-kube-api-access-9r29l\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143877 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143894 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143910 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-lib-modules\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143928 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143944 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143964 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-scripts\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143979 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.143994 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-config-data-custom\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144018 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144034 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-ceph\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144053 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144068 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144092 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144108 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-nvme\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144122 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-run\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144137 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144157 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144171 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-run\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144243 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-run\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.144916 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.145414 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-nvme\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.145451 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.145592 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.145704 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-run\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.145810 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.146487 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.146567 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.146601 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.146852 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-sys\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.146894 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.146918 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-lib-modules\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.146991 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.147135 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.147161 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.154855 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.155067 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.160408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-scripts\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.160503 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.160532 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-dev\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.161142 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.161200 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.161196 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.161414 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.164060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-ceph\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.164382 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-config-data-custom\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.164724 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.167975 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.168610 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-config-data\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.169487 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l76s\" (UniqueName: \"kubernetes.io/projected/ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047-kube-api-access-2l76s\") pod \"cinder-backup-0\" (UID: \"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047\") " pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.172448 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r29l\" (UniqueName: \"kubernetes.io/projected/a937c3c1-2bbf-4cf8-b5ac-9b55755df0af-kube-api-access-9r29l\") pod \"cinder-volume-volume1-0\" (UID: \"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af\") " pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.267416 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.345034 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.714462 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-2cfvg"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.715778 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.737932 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2cfvg"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.761542 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v42lm\" (UniqueName: \"kubernetes.io/projected/cf6e3f28-0a31-4cb0-8213-988ce169be1a-kube-api-access-v42lm\") pod \"manila-db-create-2cfvg\" (UID: \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\") " pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.761706 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6e3f28-0a31-4cb0-8213-988ce169be1a-operator-scripts\") pod \"manila-db-create-2cfvg\" (UID: \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\") " pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.810298 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.818629 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.828836 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.829041 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.829143 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kppfj" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.829595 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.836679 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-c3e6-account-create-fk69w"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.837885 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.846110 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.849290 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.862625 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.862670 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-operator-scripts\") pod \"manila-c3e6-account-create-fk69w\" (UID: \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\") " pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.862692 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb6vj\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-kube-api-access-zb6vj\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.862730 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbvgh\" (UniqueName: \"kubernetes.io/projected/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-kube-api-access-kbvgh\") pod \"manila-c3e6-account-create-fk69w\" (UID: \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\") " pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.862751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.862773 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6e3f28-0a31-4cb0-8213-988ce169be1a-operator-scripts\") pod \"manila-db-create-2cfvg\" (UID: \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\") " pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.862829 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.862854 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.865060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6e3f28-0a31-4cb0-8213-988ce169be1a-operator-scripts\") pod \"manila-db-create-2cfvg\" (UID: \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\") " pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.865132 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-ceph\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.865164 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.865271 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-logs\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.865337 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v42lm\" (UniqueName: \"kubernetes.io/projected/cf6e3f28-0a31-4cb0-8213-988ce169be1a-kube-api-access-v42lm\") pod \"manila-db-create-2cfvg\" (UID: \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\") " pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.865360 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.894435 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-c3e6-account-create-fk69w"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.897480 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v42lm\" (UniqueName: \"kubernetes.io/projected/cf6e3f28-0a31-4cb0-8213-988ce169be1a-kube-api-access-v42lm\") pod \"manila-db-create-2cfvg\" (UID: \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\") " pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.933450 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.941254 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967448 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967515 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967537 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967559 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-ceph\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967580 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-logs\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967660 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967698 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967721 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-operator-scripts\") pod \"manila-c3e6-account-create-fk69w\" (UID: \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\") " pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967739 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb6vj\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-kube-api-access-zb6vj\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.967774 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbvgh\" (UniqueName: \"kubernetes.io/projected/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-kube-api-access-kbvgh\") pod \"manila-c3e6-account-create-fk69w\" (UID: \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\") " pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.969206 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-logs\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.969653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-operator-scripts\") pod \"manila-c3e6-account-create-fk69w\" (UID: \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\") " pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.971318 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.971684 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.973967 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-config-data\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.976115 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-scripts\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.976348 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:15 crc kubenswrapper[4726]: I1123 21:02:15.977859 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.001718 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb6vj\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-kube-api-access-zb6vj\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.003566 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-ceph\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.005227 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbvgh\" (UniqueName: \"kubernetes.io/projected/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-kube-api-access-kbvgh\") pod \"manila-c3e6-account-create-fk69w\" (UID: \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\") " pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.014274 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-784b68768f-h7jf4"] Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.018040 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.024458 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.029767 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.029958 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.030077 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.030225 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-cgjv8" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.030321 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-784b68768f-h7jf4"] Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.053245 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.069022 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.069383 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.086583 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.089426 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.097045 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.097141 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.104303 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.161573 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047","Type":"ContainerStarted","Data":"ad5ae63b80ad8657ccdca2414443529e58ed50a05130953696d2ae7aff91a2e0"} Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.172455 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0281d2a-1a6f-457e-b010-f52a1d8500d7-logs\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.172502 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-config-data\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.172642 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4fmn\" (UniqueName: \"kubernetes.io/projected/e0281d2a-1a6f-457e-b010-f52a1d8500d7-kube-api-access-x4fmn\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.172689 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-scripts\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.172718 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0281d2a-1a6f-457e-b010-f52a1d8500d7-horizon-secret-key\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.181232 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.189315 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-cc8f455b9-mwcrj"] Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.197100 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.209649 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cc8f455b9-mwcrj"] Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.220483 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:16 crc kubenswrapper[4726]: E1123 21:02:16.226200 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-rv5p9 logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="ef72e5ee-5112-4956-be20-55c23f7fe4e9" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.275614 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv5p9\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-kube-api-access-rv5p9\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.275991 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276017 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276058 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276079 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276105 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4fmn\" (UniqueName: \"kubernetes.io/projected/e0281d2a-1a6f-457e-b010-f52a1d8500d7-kube-api-access-x4fmn\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276166 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276204 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-scripts\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276225 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0281d2a-1a6f-457e-b010-f52a1d8500d7-horizon-secret-key\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276247 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276309 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0281d2a-1a6f-457e-b010-f52a1d8500d7-logs\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276326 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-config-data\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.276375 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.277616 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-scripts\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.280314 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0281d2a-1a6f-457e-b010-f52a1d8500d7-logs\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.280355 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-config-data\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.331812 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4fmn\" (UniqueName: \"kubernetes.io/projected/e0281d2a-1a6f-457e-b010-f52a1d8500d7-kube-api-access-x4fmn\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.332060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0281d2a-1a6f-457e-b010-f52a1d8500d7-horizon-secret-key\") pod \"horizon-784b68768f-h7jf4\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.344339 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379308 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379360 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379408 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-scripts\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379432 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/088fec86-26a5-48f6-9d3b-453c1c6c2847-horizon-secret-key\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379459 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379504 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47mdd\" (UniqueName: \"kubernetes.io/projected/088fec86-26a5-48f6-9d3b-453c1c6c2847-kube-api-access-47mdd\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379563 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-config-data\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379588 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379619 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/088fec86-26a5-48f6-9d3b-453c1c6c2847-logs\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379669 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv5p9\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-kube-api-access-rv5p9\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379689 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379711 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379730 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.379750 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.380038 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.396348 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.399945 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.400315 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.427454 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.445893 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.463301 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.466325 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.468967 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv5p9\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-kube-api-access-rv5p9\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.493515 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-scripts\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.493750 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/088fec86-26a5-48f6-9d3b-453c1c6c2847-horizon-secret-key\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.493787 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47mdd\" (UniqueName: \"kubernetes.io/projected/088fec86-26a5-48f6-9d3b-453c1c6c2847-kube-api-access-47mdd\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.493831 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-config-data\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.500581 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/088fec86-26a5-48f6-9d3b-453c1c6c2847-logs\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.502021 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-config-data\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.505162 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/088fec86-26a5-48f6-9d3b-453c1c6c2847-logs\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.508673 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.510337 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/088fec86-26a5-48f6-9d3b-453c1c6c2847-horizon-secret-key\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.510902 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-scripts\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.518804 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47mdd\" (UniqueName: \"kubernetes.io/projected/088fec86-26a5-48f6-9d3b-453c1c6c2847-kube-api-access-47mdd\") pod \"horizon-cc8f455b9-mwcrj\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.582592 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:16 crc kubenswrapper[4726]: I1123 21:02:16.769356 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2cfvg"] Nov 23 21:02:16 crc kubenswrapper[4726]: W1123 21:02:16.892879 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf6e3f28_0a31_4cb0_8213_988ce169be1a.slice/crio-76e8cacc9c0717ea3cd1c36fd19cd83cc761523e9d365b75523092b5753661d3 WatchSource:0}: Error finding container 76e8cacc9c0717ea3cd1c36fd19cd83cc761523e9d365b75523092b5753661d3: Status 404 returned error can't find the container with id 76e8cacc9c0717ea3cd1c36fd19cd83cc761523e9d365b75523092b5753661d3 Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.055777 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.110664 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-784b68768f-h7jf4"] Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.166938 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-c3e6-account-create-fk69w"] Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.181002 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af","Type":"ContainerStarted","Data":"fa25e8c85b946fa03e1e2a3b526cbc2be7fcb72dc13385d5674c4f176811c206"} Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.183235 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.183250 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2cfvg" event={"ID":"cf6e3f28-0a31-4cb0-8213-988ce169be1a","Type":"ContainerStarted","Data":"76e8cacc9c0717ea3cd1c36fd19cd83cc761523e9d365b75523092b5753661d3"} Nov 23 21:02:17 crc kubenswrapper[4726]: W1123 21:02:17.194260 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8e3f5ad_3617_40ce_9053_a32c9d2cde1c.slice/crio-db4f8600f6e543e2ca8d7a535e428e0fca1acf799a601d716b64aefc892156f5 WatchSource:0}: Error finding container db4f8600f6e543e2ca8d7a535e428e0fca1acf799a601d716b64aefc892156f5: Status 404 returned error can't find the container with id db4f8600f6e543e2ca8d7a535e428e0fca1acf799a601d716b64aefc892156f5 Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.405538 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.429831 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cc8f455b9-mwcrj"] Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.545489 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.545721 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-internal-tls-certs\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.545773 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-combined-ca-bundle\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.545826 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-config-data\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.545894 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-ceph\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.545950 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-scripts\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.545982 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-logs\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.546007 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv5p9\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-kube-api-access-rv5p9\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.546078 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-httpd-run\") pod \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\" (UID: \"ef72e5ee-5112-4956-be20-55c23f7fe4e9\") " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.546706 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.547116 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-logs" (OuterVolumeSpecName: "logs") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.556080 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-scripts" (OuterVolumeSpecName: "scripts") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.556676 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-ceph" (OuterVolumeSpecName: "ceph") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.557967 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.558009 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.558046 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.559828 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-kube-api-access-rv5p9" (OuterVolumeSpecName: "kube-api-access-rv5p9") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "kube-api-access-rv5p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.564129 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-config-data" (OuterVolumeSpecName: "config-data") pod "ef72e5ee-5112-4956-be20-55c23f7fe4e9" (UID: "ef72e5ee-5112-4956-be20-55c23f7fe4e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648453 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648487 4726 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648498 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648509 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648517 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648525 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef72e5ee-5112-4956-be20-55c23f7fe4e9-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648532 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-logs\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648540 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv5p9\" (UniqueName: \"kubernetes.io/projected/ef72e5ee-5112-4956-be20-55c23f7fe4e9-kube-api-access-rv5p9\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.648548 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef72e5ee-5112-4956-be20-55c23f7fe4e9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.669419 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.670314 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 23 21:02:17 crc kubenswrapper[4726]: I1123 21:02:17.750608 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.235344 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8e3f5ad-3617-40ce-9053-a32c9d2cde1c" containerID="5627ac8f7118a0b69cf2b9646826d7b84145f83703783f18735d6f43c79b2899" exitCode=0 Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.235544 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-c3e6-account-create-fk69w" event={"ID":"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c","Type":"ContainerDied","Data":"5627ac8f7118a0b69cf2b9646826d7b84145f83703783f18735d6f43c79b2899"} Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.235888 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-c3e6-account-create-fk69w" event={"ID":"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c","Type":"ContainerStarted","Data":"db4f8600f6e543e2ca8d7a535e428e0fca1acf799a601d716b64aefc892156f5"} Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.245554 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9","Type":"ContainerStarted","Data":"65259ab57fd95f56e123c88738199f150a9fb19be3f936008461cda967b11a6e"} Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.250169 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cc8f455b9-mwcrj" event={"ID":"088fec86-26a5-48f6-9d3b-453c1c6c2847","Type":"ContainerStarted","Data":"00121c21913b52c6cea9d5e70958009fa3ecac5617d69e689fceaebed89fcf7f"} Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.276595 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047","Type":"ContainerStarted","Data":"e673d5a45d1de01031002420d4ecd9d267df5e020816eaaf8a4e8e1e16caf48c"} Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.276664 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047","Type":"ContainerStarted","Data":"45108aa8a10cec09ebd5e2b74bdf7ccc4461e3d7537632a6e24b5e31c5da9b27"} Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.281749 4726 generic.go:334] "Generic (PLEG): container finished" podID="cf6e3f28-0a31-4cb0-8213-988ce169be1a" containerID="ac651bac234623357112a86a3868f9d9d20f2142f4cec3b7348ab1bdf61b5435" exitCode=0 Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.281850 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2cfvg" event={"ID":"cf6e3f28-0a31-4cb0-8213-988ce169be1a","Type":"ContainerDied","Data":"ac651bac234623357112a86a3868f9d9d20f2142f4cec3b7348ab1bdf61b5435"} Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.295371 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.297229 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784b68768f-h7jf4" event={"ID":"e0281d2a-1a6f-457e-b010-f52a1d8500d7","Type":"ContainerStarted","Data":"8f18f83a7e17cecdd9c3d5b1ba3afd9a2cf82e77b9af1a6035f500457a240b05"} Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.355062 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.309435798 podStartE2EDuration="4.355036483s" podCreationTimestamp="2025-11-23 21:02:14 +0000 UTC" firstStartedPulling="2025-11-23 21:02:15.933165084 +0000 UTC m=+3244.082206040" lastFinishedPulling="2025-11-23 21:02:16.978765769 +0000 UTC m=+3245.127806725" observedRunningTime="2025-11-23 21:02:18.315025336 +0000 UTC m=+3246.464066302" watchObservedRunningTime="2025-11-23 21:02:18.355036483 +0000 UTC m=+3246.504077449" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.441913 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-784b68768f-h7jf4"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.483404 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.500085 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.541293 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5c4959b8fd-6kjm6"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.542795 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.551189 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.579967 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.581772 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.586821 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c4959b8fd-6kjm6"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.592801 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.592982 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.691325 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef72e5ee-5112-4956-be20-55c23f7fe4e9" path="/var/lib/kubelet/pods/ef72e5ee-5112-4956-be20-55c23f7fe4e9/volumes" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.692736 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.692763 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cc8f455b9-mwcrj"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703183 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-secret-key\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703228 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd63a64c-bc17-40ad-abde-de910ebbe0eb-logs\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703262 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-ceph\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703304 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-config-data\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703333 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703350 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-combined-ca-bundle\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703371 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703399 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-scripts\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703431 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703450 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-tls-certs\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703478 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703503 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-logs\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703556 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-config-data\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703590 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5q2z\" (UniqueName: \"kubernetes.io/projected/cd63a64c-bc17-40ad-abde-de910ebbe0eb-kube-api-access-x5q2z\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703637 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jpcq\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-kube-api-access-7jpcq\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.703684 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-scripts\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.785565 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:18 crc kubenswrapper[4726]: E1123 21:02:18.793703 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-7jpcq logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="79bc5930-34c6-4b82-ae50-11085c8bf084" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806293 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-config-data\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806346 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5q2z\" (UniqueName: \"kubernetes.io/projected/cd63a64c-bc17-40ad-abde-de910ebbe0eb-kube-api-access-x5q2z\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806397 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jpcq\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-kube-api-access-7jpcq\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806440 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-scripts\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806467 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-secret-key\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806484 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd63a64c-bc17-40ad-abde-de910ebbe0eb-logs\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806507 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-ceph\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806536 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-config-data\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806561 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806583 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806600 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-combined-ca-bundle\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-scripts\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806645 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806665 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-tls-certs\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806691 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.806714 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-logs\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.807221 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-logs\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.811049 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd63a64c-bc17-40ad-abde-de910ebbe0eb-logs\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.812861 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.815141 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.817805 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-scripts\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.817915 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-scripts\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.818406 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-secret-key\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.822321 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-config-data\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.826039 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-combined-ca-bundle\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.827473 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.827792 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-ceph\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.831115 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.833998 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5c65c799bb-97hbb"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.835683 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.845198 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-config-data\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.869122 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jpcq\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-kube-api-access-7jpcq\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.876797 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c65c799bb-97hbb"] Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.882522 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5q2z\" (UniqueName: \"kubernetes.io/projected/cd63a64c-bc17-40ad-abde-de910ebbe0eb-kube-api-access-x5q2z\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.897446 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-tls-certs\") pod \"horizon-5c4959b8fd-6kjm6\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.911208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.916534 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30f69dae-368c-4ccf-9d4b-e7f943d7e662-config-data\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.916599 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-combined-ca-bundle\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.916641 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-horizon-tls-certs\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.916680 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wmjx\" (UniqueName: \"kubernetes.io/projected/30f69dae-368c-4ccf-9d4b-e7f943d7e662-kube-api-access-5wmjx\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.916728 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30f69dae-368c-4ccf-9d4b-e7f943d7e662-logs\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.916804 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-horizon-secret-key\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:18 crc kubenswrapper[4726]: I1123 21:02:18.916821 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30f69dae-368c-4ccf-9d4b-e7f943d7e662-scripts\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.022884 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30f69dae-368c-4ccf-9d4b-e7f943d7e662-config-data\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.022933 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-combined-ca-bundle\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.022970 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-horizon-tls-certs\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.023005 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wmjx\" (UniqueName: \"kubernetes.io/projected/30f69dae-368c-4ccf-9d4b-e7f943d7e662-kube-api-access-5wmjx\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.023027 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30f69dae-368c-4ccf-9d4b-e7f943d7e662-logs\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.023089 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-horizon-secret-key\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.023104 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30f69dae-368c-4ccf-9d4b-e7f943d7e662-scripts\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.023771 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30f69dae-368c-4ccf-9d4b-e7f943d7e662-scripts\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.028944 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30f69dae-368c-4ccf-9d4b-e7f943d7e662-logs\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.030780 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30f69dae-368c-4ccf-9d4b-e7f943d7e662-config-data\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.033981 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-horizon-tls-certs\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.039073 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-horizon-secret-key\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.040776 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30f69dae-368c-4ccf-9d4b-e7f943d7e662-combined-ca-bundle\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.047742 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wmjx\" (UniqueName: \"kubernetes.io/projected/30f69dae-368c-4ccf-9d4b-e7f943d7e662-kube-api-access-5wmjx\") pod \"horizon-5c65c799bb-97hbb\" (UID: \"30f69dae-368c-4ccf-9d4b-e7f943d7e662\") " pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.081212 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.174948 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.363135 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af","Type":"ContainerStarted","Data":"c1fdf7ba4d72c047be8e06a462b35cc392eefe2cd171af9cd6d3899b0bc130c9"} Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.364260 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a937c3c1-2bbf-4cf8-b5ac-9b55755df0af","Type":"ContainerStarted","Data":"99af17c246add71934a8709127a80bb88a2c27ef3e9c01ebde92cf676c719a70"} Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.372658 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9","Type":"ContainerStarted","Data":"f728b316b98eab253d4cc1200ed12467cb73cb104b6b91ad00ea76ece1962017"} Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.372787 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.420400 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533059 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-config-data\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533111 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-httpd-run\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533177 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-combined-ca-bundle\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533245 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-scripts\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533275 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-internal-tls-certs\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533315 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-ceph\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533333 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-logs\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533419 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jpcq\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-kube-api-access-7jpcq\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.533491 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"79bc5930-34c6-4b82-ae50-11085c8bf084\" (UID: \"79bc5930-34c6-4b82-ae50-11085c8bf084\") " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.536392 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-logs" (OuterVolumeSpecName: "logs") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.537841 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-logs\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.541831 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.544569 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.552024 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-scripts" (OuterVolumeSpecName: "scripts") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.555045 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-config-data" (OuterVolumeSpecName: "config-data") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.555152 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-ceph" (OuterVolumeSpecName: "ceph") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.555796 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-kube-api-access-7jpcq" (OuterVolumeSpecName: "kube-api-access-7jpcq") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "kube-api-access-7jpcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.561082 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.561239 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "79bc5930-34c6-4b82-ae50-11085c8bf084" (UID: "79bc5930-34c6-4b82-ae50-11085c8bf084"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.639355 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.639388 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.639397 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79bc5930-34c6-4b82-ae50-11085c8bf084-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.639405 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.639417 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.639424 4726 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79bc5930-34c6-4b82-ae50-11085c8bf084-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.639433 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.639441 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jpcq\" (UniqueName: \"kubernetes.io/projected/79bc5930-34c6-4b82-ae50-11085c8bf084-kube-api-access-7jpcq\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.677653 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.741891 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.935550 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:19 crc kubenswrapper[4726]: I1123 21:02:19.958930 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=5.019174021 podStartE2EDuration="5.95891368s" podCreationTimestamp="2025-11-23 21:02:14 +0000 UTC" firstStartedPulling="2025-11-23 21:02:17.087197796 +0000 UTC m=+3245.236238752" lastFinishedPulling="2025-11-23 21:02:18.026937455 +0000 UTC m=+3246.175978411" observedRunningTime="2025-11-23 21:02:19.393063896 +0000 UTC m=+3247.542104852" watchObservedRunningTime="2025-11-23 21:02:19.95891368 +0000 UTC m=+3248.107954636" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.043255 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c4959b8fd-6kjm6"] Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.049505 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6e3f28-0a31-4cb0-8213-988ce169be1a-operator-scripts\") pod \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\" (UID: \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\") " Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.049698 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v42lm\" (UniqueName: \"kubernetes.io/projected/cf6e3f28-0a31-4cb0-8213-988ce169be1a-kube-api-access-v42lm\") pod \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\" (UID: \"cf6e3f28-0a31-4cb0-8213-988ce169be1a\") " Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.050243 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf6e3f28-0a31-4cb0-8213-988ce169be1a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf6e3f28-0a31-4cb0-8213-988ce169be1a" (UID: "cf6e3f28-0a31-4cb0-8213-988ce169be1a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.092210 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.103589 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c65c799bb-97hbb"] Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.103781 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf6e3f28-0a31-4cb0-8213-988ce169be1a-kube-api-access-v42lm" (OuterVolumeSpecName: "kube-api-access-v42lm") pod "cf6e3f28-0a31-4cb0-8213-988ce169be1a" (UID: "cf6e3f28-0a31-4cb0-8213-988ce169be1a"). InnerVolumeSpecName "kube-api-access-v42lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:20 crc kubenswrapper[4726]: W1123 21:02:20.143917 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30f69dae_368c_4ccf_9d4b_e7f943d7e662.slice/crio-5b8ecde8b3e7552f2355aae73bd832bfa38197d0e923c56752fcfb1ab201c16a WatchSource:0}: Error finding container 5b8ecde8b3e7552f2355aae73bd832bfa38197d0e923c56752fcfb1ab201c16a: Status 404 returned error can't find the container with id 5b8ecde8b3e7552f2355aae73bd832bfa38197d0e923c56752fcfb1ab201c16a Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.152048 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf6e3f28-0a31-4cb0-8213-988ce169be1a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.152087 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v42lm\" (UniqueName: \"kubernetes.io/projected/cf6e3f28-0a31-4cb0-8213-988ce169be1a-kube-api-access-v42lm\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.253846 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-operator-scripts\") pod \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\" (UID: \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\") " Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.254009 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbvgh\" (UniqueName: \"kubernetes.io/projected/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-kube-api-access-kbvgh\") pod \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\" (UID: \"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c\") " Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.255506 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8e3f5ad-3617-40ce-9053-a32c9d2cde1c" (UID: "a8e3f5ad-3617-40ce-9053-a32c9d2cde1c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.268005 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.275929 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-kube-api-access-kbvgh" (OuterVolumeSpecName: "kube-api-access-kbvgh") pod "a8e3f5ad-3617-40ce-9053-a32c9d2cde1c" (UID: "a8e3f5ad-3617-40ce-9053-a32c9d2cde1c"). InnerVolumeSpecName "kube-api-access-kbvgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.345636 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.356248 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbvgh\" (UniqueName: \"kubernetes.io/projected/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-kube-api-access-kbvgh\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.356286 4726 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.384316 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-c3e6-account-create-fk69w" event={"ID":"a8e3f5ad-3617-40ce-9053-a32c9d2cde1c","Type":"ContainerDied","Data":"db4f8600f6e543e2ca8d7a535e428e0fca1acf799a601d716b64aefc892156f5"} Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.384366 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db4f8600f6e543e2ca8d7a535e428e0fca1acf799a601d716b64aefc892156f5" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.384325 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-c3e6-account-create-fk69w" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.394599 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c65c799bb-97hbb" event={"ID":"30f69dae-368c-4ccf-9d4b-e7f943d7e662","Type":"ContainerStarted","Data":"5b8ecde8b3e7552f2355aae73bd832bfa38197d0e923c56752fcfb1ab201c16a"} Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.397256 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c4959b8fd-6kjm6" event={"ID":"cd63a64c-bc17-40ad-abde-de910ebbe0eb","Type":"ContainerStarted","Data":"3b857eb0eb7ac38e02fd46410a19c2ed479a5d5890a582a8441f1100b4156847"} Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.408852 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2cfvg" event={"ID":"cf6e3f28-0a31-4cb0-8213-988ce169be1a","Type":"ContainerDied","Data":"76e8cacc9c0717ea3cd1c36fd19cd83cc761523e9d365b75523092b5753661d3"} Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.408900 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76e8cacc9c0717ea3cd1c36fd19cd83cc761523e9d365b75523092b5753661d3" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.408959 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2cfvg" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.409565 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.497406 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.511460 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.517436 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:20 crc kubenswrapper[4726]: E1123 21:02:20.517862 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e3f5ad-3617-40ce-9053-a32c9d2cde1c" containerName="mariadb-account-create" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.517902 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e3f5ad-3617-40ce-9053-a32c9d2cde1c" containerName="mariadb-account-create" Nov 23 21:02:20 crc kubenswrapper[4726]: E1123 21:02:20.517928 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf6e3f28-0a31-4cb0-8213-988ce169be1a" containerName="mariadb-database-create" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.517937 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf6e3f28-0a31-4cb0-8213-988ce169be1a" containerName="mariadb-database-create" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.518234 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8e3f5ad-3617-40ce-9053-a32c9d2cde1c" containerName="mariadb-account-create" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.518281 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf6e3f28-0a31-4cb0-8213-988ce169be1a" containerName="mariadb-database-create" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.519692 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.534673 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.534997 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.540958 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.605026 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79bc5930-34c6-4b82-ae50-11085c8bf084" path="/var/lib/kubelet/pods/79bc5930-34c6-4b82-ae50-11085c8bf084/volumes" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.673581 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/97a134ee-4835-43e2-82b3-e1f19fa691cc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.673673 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.673721 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.673759 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcgqx\" (UniqueName: \"kubernetes.io/projected/97a134ee-4835-43e2-82b3-e1f19fa691cc-kube-api-access-qcgqx\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.673813 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97a134ee-4835-43e2-82b3-e1f19fa691cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.674554 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97a134ee-4835-43e2-82b3-e1f19fa691cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.674613 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.674630 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.674668 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777010 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97a134ee-4835-43e2-82b3-e1f19fa691cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777065 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777081 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777115 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777172 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/97a134ee-4835-43e2-82b3-e1f19fa691cc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777199 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777221 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777243 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcgqx\" (UniqueName: \"kubernetes.io/projected/97a134ee-4835-43e2-82b3-e1f19fa691cc-kube-api-access-qcgqx\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777268 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97a134ee-4835-43e2-82b3-e1f19fa691cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777857 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.777898 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97a134ee-4835-43e2-82b3-e1f19fa691cc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.778315 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97a134ee-4835-43e2-82b3-e1f19fa691cc-logs\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.826043 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.826516 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.827065 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/97a134ee-4835-43e2-82b3-e1f19fa691cc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.827631 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.827775 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a134ee-4835-43e2-82b3-e1f19fa691cc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.832843 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcgqx\" (UniqueName: \"kubernetes.io/projected/97a134ee-4835-43e2-82b3-e1f19fa691cc-kube-api-access-qcgqx\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:20 crc kubenswrapper[4726]: I1123 21:02:20.922018 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"97a134ee-4835-43e2-82b3-e1f19fa691cc\") " pod="openstack/glance-default-internal-api-0" Nov 23 21:02:21 crc kubenswrapper[4726]: I1123 21:02:21.172229 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:21 crc kubenswrapper[4726]: I1123 21:02:21.426128 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerName="glance-log" containerID="cri-o://f728b316b98eab253d4cc1200ed12467cb73cb104b6b91ad00ea76ece1962017" gracePeriod=30 Nov 23 21:02:21 crc kubenswrapper[4726]: I1123 21:02:21.426741 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9","Type":"ContainerStarted","Data":"57ad525891ee27ac622587528a722fd9372024ccca3d78af611fa71622cd8f0b"} Nov 23 21:02:21 crc kubenswrapper[4726]: I1123 21:02:21.427127 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerName="glance-httpd" containerID="cri-o://57ad525891ee27ac622587528a722fd9372024ccca3d78af611fa71622cd8f0b" gracePeriod=30 Nov 23 21:02:21 crc kubenswrapper[4726]: I1123 21:02:21.478365 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.47834621 podStartE2EDuration="7.47834621s" podCreationTimestamp="2025-11-23 21:02:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:02:21.452445247 +0000 UTC m=+3249.601486213" watchObservedRunningTime="2025-11-23 21:02:21.47834621 +0000 UTC m=+3249.627387166" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.456222 4726 generic.go:334] "Generic (PLEG): container finished" podID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerID="57ad525891ee27ac622587528a722fd9372024ccca3d78af611fa71622cd8f0b" exitCode=0 Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.457765 4726 generic.go:334] "Generic (PLEG): container finished" podID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerID="f728b316b98eab253d4cc1200ed12467cb73cb104b6b91ad00ea76ece1962017" exitCode=143 Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.457793 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9","Type":"ContainerDied","Data":"57ad525891ee27ac622587528a722fd9372024ccca3d78af611fa71622cd8f0b"} Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.457824 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9","Type":"ContainerDied","Data":"f728b316b98eab253d4cc1200ed12467cb73cb104b6b91ad00ea76ece1962017"} Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.568432 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.572449 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.642994 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-logs\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.643058 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-scripts\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.643117 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-config-data\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.643207 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb6vj\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-kube-api-access-zb6vj\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.643236 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-httpd-run\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.643277 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.643377 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-combined-ca-bundle\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.643397 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-ceph\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.643427 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-public-tls-certs\") pod \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\" (UID: \"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9\") " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.650382 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.663943 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-logs" (OuterVolumeSpecName: "logs") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.664948 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-kube-api-access-zb6vj" (OuterVolumeSpecName: "kube-api-access-zb6vj") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "kube-api-access-zb6vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.668155 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.682340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-ceph" (OuterVolumeSpecName: "ceph") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.696105 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-scripts" (OuterVolumeSpecName: "scripts") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.739806 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.749602 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.749630 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.749640 4726 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.749656 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-logs\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.749665 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.749676 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb6vj\" (UniqueName: \"kubernetes.io/projected/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-kube-api-access-zb6vj\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.749684 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.759888 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.770651 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.786042 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-config-data" (OuterVolumeSpecName: "config-data") pod "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" (UID: "a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.851946 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.852165 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:22 crc kubenswrapper[4726]: I1123 21:02:22.852233 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.469169 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9","Type":"ContainerDied","Data":"65259ab57fd95f56e123c88738199f150a9fb19be3f936008461cda967b11a6e"} Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.469420 4726 scope.go:117] "RemoveContainer" containerID="57ad525891ee27ac622587528a722fd9372024ccca3d78af611fa71622cd8f0b" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.469558 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.472194 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a134ee-4835-43e2-82b3-e1f19fa691cc","Type":"ContainerStarted","Data":"3cee8b498e83d3f51575325533e5266fca5d09a37e065198314d0c07bd2c5cfd"} Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.472230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a134ee-4835-43e2-82b3-e1f19fa691cc","Type":"ContainerStarted","Data":"f28c641a98e7f1c0e196b11d37e20cd0f49bd2a092aad2009d6e778bb7f91bc1"} Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.536762 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.547507 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.555325 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:23 crc kubenswrapper[4726]: E1123 21:02:23.555843 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerName="glance-log" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.555904 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerName="glance-log" Nov 23 21:02:23 crc kubenswrapper[4726]: E1123 21:02:23.555920 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerName="glance-httpd" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.555927 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerName="glance-httpd" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.556158 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerName="glance-log" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.556185 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" containerName="glance-httpd" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.562042 4726 scope.go:117] "RemoveContainer" containerID="f728b316b98eab253d4cc1200ed12467cb73cb104b6b91ad00ea76ece1962017" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.562219 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.565554 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.565997 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.579345 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685509 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685555 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-ceph\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685632 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685659 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685675 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685708 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685735 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-logs\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685754 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.685778 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5m2v\" (UniqueName: \"kubernetes.io/projected/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-kube-api-access-d5m2v\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787571 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-ceph\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787650 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787669 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787699 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787728 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-logs\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.787773 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5m2v\" (UniqueName: \"kubernetes.io/projected/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-kube-api-access-d5m2v\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.788599 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-logs\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.788818 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.790391 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.795847 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-scripts\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.804008 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-config-data\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.826460 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.827077 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.827113 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-ceph\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.829587 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5m2v\" (UniqueName: \"kubernetes.io/projected/e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb-kube-api-access-d5m2v\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.850385 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb\") " pod="openstack/glance-default-external-api-0" Nov 23 21:02:23 crc kubenswrapper[4726]: I1123 21:02:23.915135 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 23 21:02:24 crc kubenswrapper[4726]: I1123 21:02:24.494348 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a134ee-4835-43e2-82b3-e1f19fa691cc","Type":"ContainerStarted","Data":"d8fd927e32b22618c46f1f535e5c92b05c8eb6085490652a24515525edf9b30b"} Nov 23 21:02:24 crc kubenswrapper[4726]: I1123 21:02:24.524995 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.524979167 podStartE2EDuration="4.524979167s" podCreationTimestamp="2025-11-23 21:02:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:02:24.512374985 +0000 UTC m=+3252.661415941" watchObservedRunningTime="2025-11-23 21:02:24.524979167 +0000 UTC m=+3252.674020123" Nov 23 21:02:24 crc kubenswrapper[4726]: I1123 21:02:24.612239 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9" path="/var/lib/kubelet/pods/a0ea4de1-dd7a-4aaa-8c8a-6d7deca622d9/volumes" Nov 23 21:02:25 crc kubenswrapper[4726]: I1123 21:02:25.479258 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 23 21:02:25 crc kubenswrapper[4726]: I1123 21:02:25.599530 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.217602 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-6bxcm"] Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.218950 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.220825 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-ngbng" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.221099 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.233916 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-config-data\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.233967 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxb4g\" (UniqueName: \"kubernetes.io/projected/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-kube-api-access-nxb4g\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.234042 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-combined-ca-bundle\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.234088 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-job-config-data\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.239577 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-6bxcm"] Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.335508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-combined-ca-bundle\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.335589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-job-config-data\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.335661 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-config-data\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.335686 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxb4g\" (UniqueName: \"kubernetes.io/projected/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-kube-api-access-nxb4g\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.341951 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-combined-ca-bundle\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.343540 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-job-config-data\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.344647 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-config-data\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.360092 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxb4g\" (UniqueName: \"kubernetes.io/projected/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-kube-api-access-nxb4g\") pod \"manila-db-sync-6bxcm\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:26 crc kubenswrapper[4726]: I1123 21:02:26.538881 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.014219 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-6bxcm"] Nov 23 21:02:31 crc kubenswrapper[4726]: W1123 21:02:31.025720 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d583328_dc5f_49fd_b72d_c7dfbcbf1a3b.slice/crio-37b6d701364b0438eadfdc039a72dbdf80512990facb3b4b28343b894e3c2dc5 WatchSource:0}: Error finding container 37b6d701364b0438eadfdc039a72dbdf80512990facb3b4b28343b894e3c2dc5: Status 404 returned error can't find the container with id 37b6d701364b0438eadfdc039a72dbdf80512990facb3b4b28343b894e3c2dc5 Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.115779 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 23 21:02:31 crc kubenswrapper[4726]: W1123 21:02:31.122104 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4b7bb46_74e2_4aa0_8924_c8e325ad8cfb.slice/crio-93145c041a6902747a08b55161932b116ef511c87d55685729f8758bad5848c6 WatchSource:0}: Error finding container 93145c041a6902747a08b55161932b116ef511c87d55685729f8758bad5848c6: Status 404 returned error can't find the container with id 93145c041a6902747a08b55161932b116ef511c87d55685729f8758bad5848c6 Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.172641 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.172690 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.215016 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.252981 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.586098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb","Type":"ContainerStarted","Data":"93145c041a6902747a08b55161932b116ef511c87d55685729f8758bad5848c6"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.587951 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c65c799bb-97hbb" event={"ID":"30f69dae-368c-4ccf-9d4b-e7f943d7e662","Type":"ContainerStarted","Data":"b04621380fb9c0d53569cfe404ce95243271dd3240dc07231b8d51b29cd11920"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.587976 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c65c799bb-97hbb" event={"ID":"30f69dae-368c-4ccf-9d4b-e7f943d7e662","Type":"ContainerStarted","Data":"e2978518b34a3fe731ebf344e92fa2986e703389e43ae5eb721f7b3803324ec6"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.593297 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cc8f455b9-mwcrj" event={"ID":"088fec86-26a5-48f6-9d3b-453c1c6c2847","Type":"ContainerStarted","Data":"b4e3a98c4eca20e23837ae13c71fed0b5631b3241579af74739be642b28a200a"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.593338 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cc8f455b9-mwcrj" event={"ID":"088fec86-26a5-48f6-9d3b-453c1c6c2847","Type":"ContainerStarted","Data":"a255b22442ef792e115ddeb623ff5c601dfd3448f5c8b8eefc8b8bb74b582ec9"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.593498 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-cc8f455b9-mwcrj" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerName="horizon-log" containerID="cri-o://a255b22442ef792e115ddeb623ff5c601dfd3448f5c8b8eefc8b8bb74b582ec9" gracePeriod=30 Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.593581 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-cc8f455b9-mwcrj" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerName="horizon" containerID="cri-o://b4e3a98c4eca20e23837ae13c71fed0b5631b3241579af74739be642b28a200a" gracePeriod=30 Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.608602 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c4959b8fd-6kjm6" event={"ID":"cd63a64c-bc17-40ad-abde-de910ebbe0eb","Type":"ContainerStarted","Data":"ce24014cd6ad57c528c327d89f0076a660767b4d3896f6e542a9195d1ce95990"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.608645 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c4959b8fd-6kjm6" event={"ID":"cd63a64c-bc17-40ad-abde-de910ebbe0eb","Type":"ContainerStarted","Data":"be151bd7b8569f91e73ee2e0e412fce84692f7f3c3868f14f8343c70bd195ee0"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.620742 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5c65c799bb-97hbb" podStartSLOduration=3.319481936 podStartE2EDuration="13.620720192s" podCreationTimestamp="2025-11-23 21:02:18 +0000 UTC" firstStartedPulling="2025-11-23 21:02:20.161675109 +0000 UTC m=+3248.310716065" lastFinishedPulling="2025-11-23 21:02:30.462913365 +0000 UTC m=+3258.611954321" observedRunningTime="2025-11-23 21:02:31.616113593 +0000 UTC m=+3259.765154559" watchObservedRunningTime="2025-11-23 21:02:31.620720192 +0000 UTC m=+3259.769761148" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.622411 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784b68768f-h7jf4" event={"ID":"e0281d2a-1a6f-457e-b010-f52a1d8500d7","Type":"ContainerStarted","Data":"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.622452 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784b68768f-h7jf4" event={"ID":"e0281d2a-1a6f-457e-b010-f52a1d8500d7","Type":"ContainerStarted","Data":"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.622567 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-784b68768f-h7jf4" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerName="horizon-log" containerID="cri-o://60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec" gracePeriod=30 Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.622811 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-784b68768f-h7jf4" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerName="horizon" containerID="cri-o://c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7" gracePeriod=30 Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.627963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-6bxcm" event={"ID":"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b","Type":"ContainerStarted","Data":"37b6d701364b0438eadfdc039a72dbdf80512990facb3b4b28343b894e3c2dc5"} Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.629772 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.629795 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.679773 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-cc8f455b9-mwcrj" podStartSLOduration=2.716331226 podStartE2EDuration="15.679754429s" podCreationTimestamp="2025-11-23 21:02:16 +0000 UTC" firstStartedPulling="2025-11-23 21:02:17.453947681 +0000 UTC m=+3245.602988637" lastFinishedPulling="2025-11-23 21:02:30.417370854 +0000 UTC m=+3258.566411840" observedRunningTime="2025-11-23 21:02:31.676177649 +0000 UTC m=+3259.825218595" watchObservedRunningTime="2025-11-23 21:02:31.679754429 +0000 UTC m=+3259.828795385" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.728241 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5c4959b8fd-6kjm6" podStartSLOduration=3.375321544 podStartE2EDuration="13.728222952s" podCreationTimestamp="2025-11-23 21:02:18 +0000 UTC" firstStartedPulling="2025-11-23 21:02:20.068658723 +0000 UTC m=+3248.217699679" lastFinishedPulling="2025-11-23 21:02:30.421560131 +0000 UTC m=+3258.570601087" observedRunningTime="2025-11-23 21:02:31.720788115 +0000 UTC m=+3259.869829061" watchObservedRunningTime="2025-11-23 21:02:31.728222952 +0000 UTC m=+3259.877263908" Nov 23 21:02:31 crc kubenswrapper[4726]: I1123 21:02:31.762157 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-784b68768f-h7jf4" podStartSLOduration=3.53908338 podStartE2EDuration="16.762138438s" podCreationTimestamp="2025-11-23 21:02:15 +0000 UTC" firstStartedPulling="2025-11-23 21:02:17.194217573 +0000 UTC m=+3245.343258529" lastFinishedPulling="2025-11-23 21:02:30.417272631 +0000 UTC m=+3258.566313587" observedRunningTime="2025-11-23 21:02:31.754932057 +0000 UTC m=+3259.903973023" watchObservedRunningTime="2025-11-23 21:02:31.762138438 +0000 UTC m=+3259.911179394" Nov 23 21:02:32 crc kubenswrapper[4726]: I1123 21:02:32.679163 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb","Type":"ContainerStarted","Data":"da310ce0b6c3c949504c0dc366b9f3c24589d830b0f68414d684b85a75511c12"} Nov 23 21:02:32 crc kubenswrapper[4726]: I1123 21:02:32.679907 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb","Type":"ContainerStarted","Data":"b4cbced4438763b20a7fc6f416ed5909c02ea8c7f302f47bee151abc7993c767"} Nov 23 21:02:32 crc kubenswrapper[4726]: I1123 21:02:32.711499 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.711468706 podStartE2EDuration="9.711468706s" podCreationTimestamp="2025-11-23 21:02:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:02:32.709403388 +0000 UTC m=+3260.858444354" watchObservedRunningTime="2025-11-23 21:02:32.711468706 +0000 UTC m=+3260.860509662" Nov 23 21:02:33 crc kubenswrapper[4726]: I1123 21:02:33.686849 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 21:02:33 crc kubenswrapper[4726]: I1123 21:02:33.687743 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 21:02:33 crc kubenswrapper[4726]: I1123 21:02:33.915941 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 21:02:33 crc kubenswrapper[4726]: I1123 21:02:33.916011 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 23 21:02:33 crc kubenswrapper[4726]: I1123 21:02:33.974239 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 21:02:33 crc kubenswrapper[4726]: I1123 21:02:33.982029 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 23 21:02:34 crc kubenswrapper[4726]: I1123 21:02:34.703968 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 21:02:34 crc kubenswrapper[4726]: I1123 21:02:34.704430 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 23 21:02:36 crc kubenswrapper[4726]: I1123 21:02:36.347543 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:02:36 crc kubenswrapper[4726]: I1123 21:02:36.585046 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:02:36 crc kubenswrapper[4726]: I1123 21:02:36.982658 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:36 crc kubenswrapper[4726]: I1123 21:02:36.982754 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 23 21:02:37 crc kubenswrapper[4726]: I1123 21:02:37.051129 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 23 21:02:39 crc kubenswrapper[4726]: I1123 21:02:39.081784 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:39 crc kubenswrapper[4726]: I1123 21:02:39.082368 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:02:39 crc kubenswrapper[4726]: I1123 21:02:39.176347 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:39 crc kubenswrapper[4726]: I1123 21:02:39.176481 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:02:39 crc kubenswrapper[4726]: I1123 21:02:39.749495 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-6bxcm" event={"ID":"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b","Type":"ContainerStarted","Data":"ed3639f6be29fd4e8cb8338756f5c41b31bb7a8f749a9a72c2160cd2151346c1"} Nov 23 21:02:39 crc kubenswrapper[4726]: I1123 21:02:39.763931 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-6bxcm" podStartSLOduration=6.045135466 podStartE2EDuration="13.76387316s" podCreationTimestamp="2025-11-23 21:02:26 +0000 UTC" firstStartedPulling="2025-11-23 21:02:31.027951286 +0000 UTC m=+3259.176992242" lastFinishedPulling="2025-11-23 21:02:38.74668898 +0000 UTC m=+3266.895729936" observedRunningTime="2025-11-23 21:02:39.762245755 +0000 UTC m=+3267.911286711" watchObservedRunningTime="2025-11-23 21:02:39.76387316 +0000 UTC m=+3267.912914116" Nov 23 21:02:40 crc kubenswrapper[4726]: I1123 21:02:40.000635 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 21:02:42 crc kubenswrapper[4726]: I1123 21:02:42.178425 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 23 21:02:49 crc kubenswrapper[4726]: I1123 21:02:49.084239 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.243:8443: connect: connection refused" Nov 23 21:02:49 crc kubenswrapper[4726]: I1123 21:02:49.177246 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c65c799bb-97hbb" podUID="30f69dae-368c-4ccf-9d4b-e7f943d7e662" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.245:8443: connect: connection refused" Nov 23 21:02:49 crc kubenswrapper[4726]: I1123 21:02:49.832331 4726 generic.go:334] "Generic (PLEG): container finished" podID="4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" containerID="ed3639f6be29fd4e8cb8338756f5c41b31bb7a8f749a9a72c2160cd2151346c1" exitCode=0 Nov 23 21:02:49 crc kubenswrapper[4726]: I1123 21:02:49.832385 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-6bxcm" event={"ID":"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b","Type":"ContainerDied","Data":"ed3639f6be29fd4e8cb8338756f5c41b31bb7a8f749a9a72c2160cd2151346c1"} Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.765956 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.827389 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-job-config-data\") pod \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.827448 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxb4g\" (UniqueName: \"kubernetes.io/projected/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-kube-api-access-nxb4g\") pod \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.827487 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-combined-ca-bundle\") pod \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.827507 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-config-data\") pod \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\" (UID: \"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b\") " Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.834150 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-kube-api-access-nxb4g" (OuterVolumeSpecName: "kube-api-access-nxb4g") pod "4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" (UID: "4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b"). InnerVolumeSpecName "kube-api-access-nxb4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.837670 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" (UID: "4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.841645 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-config-data" (OuterVolumeSpecName: "config-data") pod "4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" (UID: "4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.854742 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-6bxcm" event={"ID":"4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b","Type":"ContainerDied","Data":"37b6d701364b0438eadfdc039a72dbdf80512990facb3b4b28343b894e3c2dc5"} Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.854782 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37b6d701364b0438eadfdc039a72dbdf80512990facb3b4b28343b894e3c2dc5" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.854840 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-6bxcm" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.866649 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" (UID: "4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.930807 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.930846 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.930858 4726 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:51 crc kubenswrapper[4726]: I1123 21:02:51.930888 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxb4g\" (UniqueName: \"kubernetes.io/projected/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b-kube-api-access-nxb4g\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.081921 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:02:52 crc kubenswrapper[4726]: E1123 21:02:52.094494 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" containerName="manila-db-sync" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.094533 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" containerName="manila-db-sync" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.094742 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" containerName="manila-db-sync" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.095736 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.099629 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.099836 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.121669 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.135291 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-scripts\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.135377 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.135401 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.135456 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c3f104ff-1e92-48e0-a3c8-7ec06184f578-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.135501 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p667h\" (UniqueName: \"kubernetes.io/projected/c3f104ff-1e92-48e0-a3c8-7ec06184f578-kube-api-access-p667h\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.135518 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.192967 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.202662 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.207848 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.233262 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.236767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.236811 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4q8c\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-kube-api-access-b4q8c\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.236827 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-ceph\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.236881 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.236903 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.236942 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.236970 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.236987 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c3f104ff-1e92-48e0-a3c8-7ec06184f578-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.237007 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.237037 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.237059 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p667h\" (UniqueName: \"kubernetes.io/projected/c3f104ff-1e92-48e0-a3c8-7ec06184f578-kube-api-access-p667h\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.237075 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.237089 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-scripts\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.237114 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-scripts\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.237986 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c3f104ff-1e92-48e0-a3c8-7ec06184f578-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.256943 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.285929 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.286519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.292420 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-scripts\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.324719 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p667h\" (UniqueName: \"kubernetes.io/projected/c3f104ff-1e92-48e0-a3c8-7ec06184f578-kube-api-access-p667h\") pod \"manila-scheduler-0\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.329599 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ddf789dd7-wznxr"] Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.338755 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-ceph\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.342046 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.345476 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-ceph\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.347459 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4q8c\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-kube-api-access-b4q8c\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.347907 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.347995 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.348037 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.348104 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.348159 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-scripts\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.348252 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.348474 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.349624 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.354719 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.358283 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ddf789dd7-wznxr"] Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.360339 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.362922 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.365713 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-scripts\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.389575 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4q8c\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-kube-api-access-b4q8c\") pod \"manila-share-share1-0\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.437607 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.453356 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqcqt\" (UniqueName: \"kubernetes.io/projected/6c62393d-31e9-4a24-9976-ffcf175d1443-kube-api-access-lqcqt\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.453685 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-config\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.453713 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-ovsdbserver-sb\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.453809 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-dns-svc\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.453834 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-openstack-edpm-ipam\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.453850 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-ovsdbserver-nb\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.461809 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.463521 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.477112 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.479595 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.533445 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.562976 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563048 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-dns-svc\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563077 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-openstack-edpm-ipam\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563097 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-ovsdbserver-nb\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data-custom\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563179 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqcqt\" (UniqueName: \"kubernetes.io/projected/6c62393d-31e9-4a24-9976-ffcf175d1443-kube-api-access-lqcqt\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563206 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-config\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563221 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d474804b-2f10-4994-9ce0-6c9a15aee2b1-logs\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563238 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-scripts\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563253 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563275 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-ovsdbserver-sb\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563298 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24bvx\" (UniqueName: \"kubernetes.io/projected/d474804b-2f10-4994-9ce0-6c9a15aee2b1-kube-api-access-24bvx\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.563316 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d474804b-2f10-4994-9ce0-6c9a15aee2b1-etc-machine-id\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.564362 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-dns-svc\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.564536 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-config\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.567053 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-openstack-edpm-ipam\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.577271 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-ovsdbserver-sb\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.579988 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c62393d-31e9-4a24-9976-ffcf175d1443-ovsdbserver-nb\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.583424 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqcqt\" (UniqueName: \"kubernetes.io/projected/6c62393d-31e9-4a24-9976-ffcf175d1443-kube-api-access-lqcqt\") pod \"dnsmasq-dns-7ddf789dd7-wznxr\" (UID: \"6c62393d-31e9-4a24-9976-ffcf175d1443\") " pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.667552 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data-custom\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.667721 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d474804b-2f10-4994-9ce0-6c9a15aee2b1-logs\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.667756 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.667776 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-scripts\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.667841 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24bvx\" (UniqueName: \"kubernetes.io/projected/d474804b-2f10-4994-9ce0-6c9a15aee2b1-kube-api-access-24bvx\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.667900 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d474804b-2f10-4994-9ce0-6c9a15aee2b1-etc-machine-id\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.668008 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.670311 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d474804b-2f10-4994-9ce0-6c9a15aee2b1-logs\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.677095 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d474804b-2f10-4994-9ce0-6c9a15aee2b1-etc-machine-id\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.679699 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.685682 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-scripts\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.696978 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.698280 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.709081 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24bvx\" (UniqueName: \"kubernetes.io/projected/d474804b-2f10-4994-9ce0-6c9a15aee2b1-kube-api-access-24bvx\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.715309 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data-custom\") pod \"manila-api-0\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " pod="openstack/manila-api-0" Nov 23 21:02:52 crc kubenswrapper[4726]: I1123 21:02:52.819589 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 23 21:02:53 crc kubenswrapper[4726]: I1123 21:02:53.471465 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:02:53 crc kubenswrapper[4726]: I1123 21:02:53.849636 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ddf789dd7-wznxr"] Nov 23 21:02:53 crc kubenswrapper[4726]: I1123 21:02:53.928959 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" event={"ID":"6c62393d-31e9-4a24-9976-ffcf175d1443","Type":"ContainerStarted","Data":"bdf06c5f94882afbaf633105fb2ee0832e9ac793661a540568d13ac7222f9124"} Nov 23 21:02:53 crc kubenswrapper[4726]: I1123 21:02:53.937875 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c3f104ff-1e92-48e0-a3c8-7ec06184f578","Type":"ContainerStarted","Data":"bafe90b33ca8f8c6b265879c0290bc56d0900013b580fa7788f47d1d209bc385"} Nov 23 21:02:53 crc kubenswrapper[4726]: I1123 21:02:53.970464 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:02:54 crc kubenswrapper[4726]: I1123 21:02:54.951808 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1cdd8341-a77e-4f3c-9c2e-e56125a6162c","Type":"ContainerStarted","Data":"c2be0777839febc93886ece1b47a147fee2b9dfab9a4ffbdab002f5c58fb767a"} Nov 23 21:02:54 crc kubenswrapper[4726]: I1123 21:02:54.958554 4726 generic.go:334] "Generic (PLEG): container finished" podID="6c62393d-31e9-4a24-9976-ffcf175d1443" containerID="960159a7948464e755b296b03006ac042e67a53d30e365c25e3d0a0b489ef793" exitCode=0 Nov 23 21:02:54 crc kubenswrapper[4726]: I1123 21:02:54.958599 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" event={"ID":"6c62393d-31e9-4a24-9976-ffcf175d1443","Type":"ContainerDied","Data":"960159a7948464e755b296b03006ac042e67a53d30e365c25e3d0a0b489ef793"} Nov 23 21:02:54 crc kubenswrapper[4726]: I1123 21:02:54.989595 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 23 21:02:55 crc kubenswrapper[4726]: I1123 21:02:55.969983 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c3f104ff-1e92-48e0-a3c8-7ec06184f578","Type":"ContainerStarted","Data":"fb85c536faa2e9eb2b4f98bedfc281a76086eaf19cf9a9260d7b9b84775c5779"} Nov 23 21:02:55 crc kubenswrapper[4726]: I1123 21:02:55.975366 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" event={"ID":"6c62393d-31e9-4a24-9976-ffcf175d1443","Type":"ContainerStarted","Data":"cda2222495443e205ef5db440d2cf780ba70f6596a6ce9a235d63c7fcf242d2f"} Nov 23 21:02:55 crc kubenswrapper[4726]: I1123 21:02:55.975543 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:02:55 crc kubenswrapper[4726]: I1123 21:02:55.983989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d474804b-2f10-4994-9ce0-6c9a15aee2b1","Type":"ContainerStarted","Data":"b1622493f93ea5e3aab56e9db5d8a49e6a13ecb5c686f501a1de9f1bb8924cf2"} Nov 23 21:02:56 crc kubenswrapper[4726]: I1123 21:02:56.023654 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" podStartSLOduration=4.0236353 podStartE2EDuration="4.0236353s" podCreationTimestamp="2025-11-23 21:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:02:56.000435332 +0000 UTC m=+3284.149476298" watchObservedRunningTime="2025-11-23 21:02:56.0236353 +0000 UTC m=+3284.172676256" Nov 23 21:02:56 crc kubenswrapper[4726]: I1123 21:02:56.948327 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 23 21:02:56 crc kubenswrapper[4726]: I1123 21:02:56.997227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c3f104ff-1e92-48e0-a3c8-7ec06184f578","Type":"ContainerStarted","Data":"b261223dc49063d972c9c3d098c1fc66516d9fd957e1b81a41cea2d352b250aa"} Nov 23 21:02:57 crc kubenswrapper[4726]: I1123 21:02:57.000650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d474804b-2f10-4994-9ce0-6c9a15aee2b1","Type":"ContainerStarted","Data":"8c18edecc8a6c983d4a337b1f1ead199ea7b235e4992b6f6eccc6c91ddeef922"} Nov 23 21:02:57 crc kubenswrapper[4726]: I1123 21:02:57.000686 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 23 21:02:57 crc kubenswrapper[4726]: I1123 21:02:57.000696 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d474804b-2f10-4994-9ce0-6c9a15aee2b1","Type":"ContainerStarted","Data":"cb1aa9060bbe1832d9790d85beeb36b28fc3a098686ed8fd21fc16b4f3b05de0"} Nov 23 21:02:57 crc kubenswrapper[4726]: I1123 21:02:57.025895 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.265545411 podStartE2EDuration="5.025849533s" podCreationTimestamp="2025-11-23 21:02:52 +0000 UTC" firstStartedPulling="2025-11-23 21:02:53.546158118 +0000 UTC m=+3281.695199074" lastFinishedPulling="2025-11-23 21:02:54.30646224 +0000 UTC m=+3282.455503196" observedRunningTime="2025-11-23 21:02:57.017000305 +0000 UTC m=+3285.166041271" watchObservedRunningTime="2025-11-23 21:02:57.025849533 +0000 UTC m=+3285.174890489" Nov 23 21:02:57 crc kubenswrapper[4726]: I1123 21:02:57.041682 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=5.041651514 podStartE2EDuration="5.041651514s" podCreationTimestamp="2025-11-23 21:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:02:57.040034408 +0000 UTC m=+3285.189075384" watchObservedRunningTime="2025-11-23 21:02:57.041651514 +0000 UTC m=+3285.190692470" Nov 23 21:02:58 crc kubenswrapper[4726]: I1123 21:02:58.008746 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerName="manila-api-log" containerID="cri-o://cb1aa9060bbe1832d9790d85beeb36b28fc3a098686ed8fd21fc16b4f3b05de0" gracePeriod=30 Nov 23 21:02:58 crc kubenswrapper[4726]: I1123 21:02:58.009053 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerName="manila-api" containerID="cri-o://8c18edecc8a6c983d4a337b1f1ead199ea7b235e4992b6f6eccc6c91ddeef922" gracePeriod=30 Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.068998 4726 generic.go:334] "Generic (PLEG): container finished" podID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerID="8c18edecc8a6c983d4a337b1f1ead199ea7b235e4992b6f6eccc6c91ddeef922" exitCode=0 Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.069462 4726 generic.go:334] "Generic (PLEG): container finished" podID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerID="cb1aa9060bbe1832d9790d85beeb36b28fc3a098686ed8fd21fc16b4f3b05de0" exitCode=143 Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.069509 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d474804b-2f10-4994-9ce0-6c9a15aee2b1","Type":"ContainerDied","Data":"8c18edecc8a6c983d4a337b1f1ead199ea7b235e4992b6f6eccc6c91ddeef922"} Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.069564 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d474804b-2f10-4994-9ce0-6c9a15aee2b1","Type":"ContainerDied","Data":"cb1aa9060bbe1832d9790d85beeb36b28fc3a098686ed8fd21fc16b4f3b05de0"} Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.069576 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"d474804b-2f10-4994-9ce0-6c9a15aee2b1","Type":"ContainerDied","Data":"b1622493f93ea5e3aab56e9db5d8a49e6a13ecb5c686f501a1de9f1bb8924cf2"} Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.069628 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1622493f93ea5e3aab56e9db5d8a49e6a13ecb5c686f501a1de9f1bb8924cf2" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.082640 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.243:8443: connect: connection refused" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.143171 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.176270 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c65c799bb-97hbb" podUID="30f69dae-368c-4ccf-9d4b-e7f943d7e662" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.245:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.245:8443: connect: connection refused" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.242475 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d474804b-2f10-4994-9ce0-6c9a15aee2b1-etc-machine-id\") pod \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.242544 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-combined-ca-bundle\") pod \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.242590 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data-custom\") pod \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.242646 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24bvx\" (UniqueName: \"kubernetes.io/projected/d474804b-2f10-4994-9ce0-6c9a15aee2b1-kube-api-access-24bvx\") pod \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.242741 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d474804b-2f10-4994-9ce0-6c9a15aee2b1-logs\") pod \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.242914 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data\") pod \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.242960 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-scripts\") pod \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\" (UID: \"d474804b-2f10-4994-9ce0-6c9a15aee2b1\") " Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.245248 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d474804b-2f10-4994-9ce0-6c9a15aee2b1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d474804b-2f10-4994-9ce0-6c9a15aee2b1" (UID: "d474804b-2f10-4994-9ce0-6c9a15aee2b1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.254973 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d474804b-2f10-4994-9ce0-6c9a15aee2b1" (UID: "d474804b-2f10-4994-9ce0-6c9a15aee2b1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.258296 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d474804b-2f10-4994-9ce0-6c9a15aee2b1-logs" (OuterVolumeSpecName: "logs") pod "d474804b-2f10-4994-9ce0-6c9a15aee2b1" (UID: "d474804b-2f10-4994-9ce0-6c9a15aee2b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.263490 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d474804b-2f10-4994-9ce0-6c9a15aee2b1-kube-api-access-24bvx" (OuterVolumeSpecName: "kube-api-access-24bvx") pod "d474804b-2f10-4994-9ce0-6c9a15aee2b1" (UID: "d474804b-2f10-4994-9ce0-6c9a15aee2b1"). InnerVolumeSpecName "kube-api-access-24bvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.275084 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-scripts" (OuterVolumeSpecName: "scripts") pod "d474804b-2f10-4994-9ce0-6c9a15aee2b1" (UID: "d474804b-2f10-4994-9ce0-6c9a15aee2b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.337061 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d474804b-2f10-4994-9ce0-6c9a15aee2b1" (UID: "d474804b-2f10-4994-9ce0-6c9a15aee2b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.347166 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d474804b-2f10-4994-9ce0-6c9a15aee2b1-logs\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.347200 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.347211 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d474804b-2f10-4994-9ce0-6c9a15aee2b1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.347223 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.347231 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.347241 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24bvx\" (UniqueName: \"kubernetes.io/projected/d474804b-2f10-4994-9ce0-6c9a15aee2b1-kube-api-access-24bvx\") on node \"crc\" DevicePath \"\"" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.357974 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data" (OuterVolumeSpecName: "config-data") pod "d474804b-2f10-4994-9ce0-6c9a15aee2b1" (UID: "d474804b-2f10-4994-9ce0-6c9a15aee2b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:02:59 crc kubenswrapper[4726]: I1123 21:02:59.448972 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d474804b-2f10-4994-9ce0-6c9a15aee2b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.076467 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.111214 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.121590 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.153961 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 23 21:03:00 crc kubenswrapper[4726]: E1123 21:03:00.156777 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerName="manila-api-log" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.170784 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerName="manila-api-log" Nov 23 21:03:00 crc kubenswrapper[4726]: E1123 21:03:00.171018 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerName="manila-api" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.171132 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerName="manila-api" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.176148 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerName="manila-api-log" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.176198 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" containerName="manila-api" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.193259 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.196470 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.201502 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.202348 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.202901 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.275976 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-scripts\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.276048 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-config-data\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.276082 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s58bp\" (UniqueName: \"kubernetes.io/projected/f82bda1a-eb9c-4e50-9700-85c8f6a55336-kube-api-access-s58bp\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.276130 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f82bda1a-eb9c-4e50-9700-85c8f6a55336-etc-machine-id\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.276216 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-public-tls-certs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.276242 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.276274 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-config-data-custom\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.276298 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f82bda1a-eb9c-4e50-9700-85c8f6a55336-logs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.292433 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.393982 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-config-data\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.394020 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s58bp\" (UniqueName: \"kubernetes.io/projected/f82bda1a-eb9c-4e50-9700-85c8f6a55336-kube-api-access-s58bp\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.394085 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f82bda1a-eb9c-4e50-9700-85c8f6a55336-etc-machine-id\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.394149 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-public-tls-certs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.394173 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.394216 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-config-data-custom\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.394233 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f82bda1a-eb9c-4e50-9700-85c8f6a55336-logs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.394286 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.394328 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-scripts\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.399345 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f82bda1a-eb9c-4e50-9700-85c8f6a55336-logs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.399706 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-scripts\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.399805 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-public-tls-certs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.404652 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f82bda1a-eb9c-4e50-9700-85c8f6a55336-etc-machine-id\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.408451 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-config-data-custom\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.425518 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.430401 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.431262 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f82bda1a-eb9c-4e50-9700-85c8f6a55336-config-data\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.436373 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s58bp\" (UniqueName: \"kubernetes.io/projected/f82bda1a-eb9c-4e50-9700-85c8f6a55336-kube-api-access-s58bp\") pod \"manila-api-0\" (UID: \"f82bda1a-eb9c-4e50-9700-85c8f6a55336\") " pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.517998 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 23 21:03:00 crc kubenswrapper[4726]: I1123 21:03:00.614978 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d474804b-2f10-4994-9ce0-6c9a15aee2b1" path="/var/lib/kubelet/pods/d474804b-2f10-4994-9ce0-6c9a15aee2b1/volumes" Nov 23 21:03:01 crc kubenswrapper[4726]: I1123 21:03:01.192195 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 23 21:03:01 crc kubenswrapper[4726]: W1123 21:03:01.211062 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf82bda1a_eb9c_4e50_9700_85c8f6a55336.slice/crio-8570a955d0588ff352b20fd07a66b3567c0f8e6f12656d3a9c910d7f2dadba36 WatchSource:0}: Error finding container 8570a955d0588ff352b20fd07a66b3567c0f8e6f12656d3a9c910d7f2dadba36: Status 404 returned error can't find the container with id 8570a955d0588ff352b20fd07a66b3567c0f8e6f12656d3a9c910d7f2dadba36 Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.036030 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.105514 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f82bda1a-eb9c-4e50-9700-85c8f6a55336","Type":"ContainerStarted","Data":"e9506bf3ff3ca231e800e5a84cd4f051089438c268d455b37a578439484b88d9"} Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.105554 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f82bda1a-eb9c-4e50-9700-85c8f6a55336","Type":"ContainerStarted","Data":"8570a955d0588ff352b20fd07a66b3567c0f8e6f12656d3a9c910d7f2dadba36"} Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.106949 4726 generic.go:334] "Generic (PLEG): container finished" podID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerID="b4e3a98c4eca20e23837ae13c71fed0b5631b3241579af74739be642b28a200a" exitCode=137 Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.106972 4726 generic.go:334] "Generic (PLEG): container finished" podID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerID="a255b22442ef792e115ddeb623ff5c601dfd3448f5c8b8eefc8b8bb74b582ec9" exitCode=137 Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.106998 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cc8f455b9-mwcrj" event={"ID":"088fec86-26a5-48f6-9d3b-453c1c6c2847","Type":"ContainerDied","Data":"b4e3a98c4eca20e23837ae13c71fed0b5631b3241579af74739be642b28a200a"} Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.107014 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cc8f455b9-mwcrj" event={"ID":"088fec86-26a5-48f6-9d3b-453c1c6c2847","Type":"ContainerDied","Data":"a255b22442ef792e115ddeb623ff5c601dfd3448f5c8b8eefc8b8bb74b582ec9"} Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.112462 4726 generic.go:334] "Generic (PLEG): container finished" podID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerID="c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7" exitCode=137 Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.112482 4726 generic.go:334] "Generic (PLEG): container finished" podID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerID="60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec" exitCode=137 Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.112501 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784b68768f-h7jf4" event={"ID":"e0281d2a-1a6f-457e-b010-f52a1d8500d7","Type":"ContainerDied","Data":"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7"} Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.112517 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784b68768f-h7jf4" event={"ID":"e0281d2a-1a6f-457e-b010-f52a1d8500d7","Type":"ContainerDied","Data":"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec"} Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.112528 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784b68768f-h7jf4" event={"ID":"e0281d2a-1a6f-457e-b010-f52a1d8500d7","Type":"ContainerDied","Data":"8f18f83a7e17cecdd9c3d5b1ba3afd9a2cf82e77b9af1a6035f500457a240b05"} Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.112582 4726 scope.go:117] "RemoveContainer" containerID="c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.112718 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-784b68768f-h7jf4" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.131226 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0281d2a-1a6f-457e-b010-f52a1d8500d7-horizon-secret-key\") pod \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.131292 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-scripts\") pod \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.131372 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-config-data\") pod \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.131411 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4fmn\" (UniqueName: \"kubernetes.io/projected/e0281d2a-1a6f-457e-b010-f52a1d8500d7-kube-api-access-x4fmn\") pod \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.131540 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0281d2a-1a6f-457e-b010-f52a1d8500d7-logs\") pod \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\" (UID: \"e0281d2a-1a6f-457e-b010-f52a1d8500d7\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.132842 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0281d2a-1a6f-457e-b010-f52a1d8500d7-logs" (OuterVolumeSpecName: "logs") pod "e0281d2a-1a6f-457e-b010-f52a1d8500d7" (UID: "e0281d2a-1a6f-457e-b010-f52a1d8500d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.137601 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0281d2a-1a6f-457e-b010-f52a1d8500d7-kube-api-access-x4fmn" (OuterVolumeSpecName: "kube-api-access-x4fmn") pod "e0281d2a-1a6f-457e-b010-f52a1d8500d7" (UID: "e0281d2a-1a6f-457e-b010-f52a1d8500d7"). InnerVolumeSpecName "kube-api-access-x4fmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.150747 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0281d2a-1a6f-457e-b010-f52a1d8500d7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e0281d2a-1a6f-457e-b010-f52a1d8500d7" (UID: "e0281d2a-1a6f-457e-b010-f52a1d8500d7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.202406 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-config-data" (OuterVolumeSpecName: "config-data") pod "e0281d2a-1a6f-457e-b010-f52a1d8500d7" (UID: "e0281d2a-1a6f-457e-b010-f52a1d8500d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.241447 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.241755 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4fmn\" (UniqueName: \"kubernetes.io/projected/e0281d2a-1a6f-457e-b010-f52a1d8500d7-kube-api-access-x4fmn\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.241765 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0281d2a-1a6f-457e-b010-f52a1d8500d7-logs\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.241773 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0281d2a-1a6f-457e-b010-f52a1d8500d7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.264360 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-scripts" (OuterVolumeSpecName: "scripts") pod "e0281d2a-1a6f-457e-b010-f52a1d8500d7" (UID: "e0281d2a-1a6f-457e-b010-f52a1d8500d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.344917 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0281d2a-1a6f-457e-b010-f52a1d8500d7-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.438559 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.531921 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-784b68768f-h7jf4"] Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.540641 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-784b68768f-h7jf4"] Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.607692 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" path="/var/lib/kubelet/pods/e0281d2a-1a6f-457e-b010-f52a1d8500d7/volumes" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.699064 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ddf789dd7-wznxr" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.710636 4726 scope.go:117] "RemoveContainer" containerID="60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.787354 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5db79bb959-dxv4q"] Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.787592 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" podUID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" containerName="dnsmasq-dns" containerID="cri-o://0c54114e02a79af89484aae3ce55a2532c71913989ade53e32ce3596a125d907" gracePeriod=10 Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.831905 4726 scope.go:117] "RemoveContainer" containerID="c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7" Nov 23 21:03:02 crc kubenswrapper[4726]: E1123 21:03:02.833483 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7\": container with ID starting with c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7 not found: ID does not exist" containerID="c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.833533 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7"} err="failed to get container status \"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7\": rpc error: code = NotFound desc = could not find container \"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7\": container with ID starting with c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7 not found: ID does not exist" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.833557 4726 scope.go:117] "RemoveContainer" containerID="60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec" Nov 23 21:03:02 crc kubenswrapper[4726]: E1123 21:03:02.833919 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec\": container with ID starting with 60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec not found: ID does not exist" containerID="60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.833941 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec"} err="failed to get container status \"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec\": rpc error: code = NotFound desc = could not find container \"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec\": container with ID starting with 60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec not found: ID does not exist" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.833973 4726 scope.go:117] "RemoveContainer" containerID="c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.834270 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7"} err="failed to get container status \"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7\": rpc error: code = NotFound desc = could not find container \"c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7\": container with ID starting with c84694e092254ef4e39c4a0eb94b757de141139880eba2c3fb3a795a15929ca7 not found: ID does not exist" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.834315 4726 scope.go:117] "RemoveContainer" containerID="60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.835003 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec"} err="failed to get container status \"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec\": rpc error: code = NotFound desc = could not find container \"60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec\": container with ID starting with 60e738aeae862e33528af7b5d50172cbd9c0ee8ef43ce0c0f85a87c8c956d5ec not found: ID does not exist" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.861496 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.957206 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-scripts\") pod \"088fec86-26a5-48f6-9d3b-453c1c6c2847\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.957284 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-config-data\") pod \"088fec86-26a5-48f6-9d3b-453c1c6c2847\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.957378 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/088fec86-26a5-48f6-9d3b-453c1c6c2847-logs\") pod \"088fec86-26a5-48f6-9d3b-453c1c6c2847\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.957491 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/088fec86-26a5-48f6-9d3b-453c1c6c2847-horizon-secret-key\") pod \"088fec86-26a5-48f6-9d3b-453c1c6c2847\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.957573 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47mdd\" (UniqueName: \"kubernetes.io/projected/088fec86-26a5-48f6-9d3b-453c1c6c2847-kube-api-access-47mdd\") pod \"088fec86-26a5-48f6-9d3b-453c1c6c2847\" (UID: \"088fec86-26a5-48f6-9d3b-453c1c6c2847\") " Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.959122 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/088fec86-26a5-48f6-9d3b-453c1c6c2847-logs" (OuterVolumeSpecName: "logs") pod "088fec86-26a5-48f6-9d3b-453c1c6c2847" (UID: "088fec86-26a5-48f6-9d3b-453c1c6c2847"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.967067 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/088fec86-26a5-48f6-9d3b-453c1c6c2847-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "088fec86-26a5-48f6-9d3b-453c1c6c2847" (UID: "088fec86-26a5-48f6-9d3b-453c1c6c2847"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:02 crc kubenswrapper[4726]: I1123 21:03:02.976241 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088fec86-26a5-48f6-9d3b-453c1c6c2847-kube-api-access-47mdd" (OuterVolumeSpecName: "kube-api-access-47mdd") pod "088fec86-26a5-48f6-9d3b-453c1c6c2847" (UID: "088fec86-26a5-48f6-9d3b-453c1c6c2847"). InnerVolumeSpecName "kube-api-access-47mdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.003986 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-scripts" (OuterVolumeSpecName: "scripts") pod "088fec86-26a5-48f6-9d3b-453c1c6c2847" (UID: "088fec86-26a5-48f6-9d3b-453c1c6c2847"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.005605 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-config-data" (OuterVolumeSpecName: "config-data") pod "088fec86-26a5-48f6-9d3b-453c1c6c2847" (UID: "088fec86-26a5-48f6-9d3b-453c1c6c2847"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.063109 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47mdd\" (UniqueName: \"kubernetes.io/projected/088fec86-26a5-48f6-9d3b-453c1c6c2847-kube-api-access-47mdd\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.063146 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.063156 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/088fec86-26a5-48f6-9d3b-453c1c6c2847-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.063164 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/088fec86-26a5-48f6-9d3b-453c1c6c2847-logs\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.063173 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/088fec86-26a5-48f6-9d3b-453c1c6c2847-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.131722 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cc8f455b9-mwcrj" event={"ID":"088fec86-26a5-48f6-9d3b-453c1c6c2847","Type":"ContainerDied","Data":"00121c21913b52c6cea9d5e70958009fa3ecac5617d69e689fceaebed89fcf7f"} Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.131770 4726 scope.go:117] "RemoveContainer" containerID="b4e3a98c4eca20e23837ae13c71fed0b5631b3241579af74739be642b28a200a" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.131884 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cc8f455b9-mwcrj" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.142237 4726 generic.go:334] "Generic (PLEG): container finished" podID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" containerID="0c54114e02a79af89484aae3ce55a2532c71913989ade53e32ce3596a125d907" exitCode=0 Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.142301 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" event={"ID":"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef","Type":"ContainerDied","Data":"0c54114e02a79af89484aae3ce55a2532c71913989ade53e32ce3596a125d907"} Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.156183 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f82bda1a-eb9c-4e50-9700-85c8f6a55336","Type":"ContainerStarted","Data":"02c0532188d9c337352fde229981d9d47d2b8d95a2df5ae3c146d4ee85168e2a"} Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.156414 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.178940 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cc8f455b9-mwcrj"] Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.189413 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-cc8f455b9-mwcrj"] Nov 23 21:03:03 crc kubenswrapper[4726]: I1123 21:03:03.206849 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.206827685 podStartE2EDuration="3.206827685s" podCreationTimestamp="2025-11-23 21:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:03:03.200090717 +0000 UTC m=+3291.349131673" watchObservedRunningTime="2025-11-23 21:03:03.206827685 +0000 UTC m=+3291.355868631" Nov 23 21:03:04 crc kubenswrapper[4726]: I1123 21:03:04.600158 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" path="/var/lib/kubelet/pods/088fec86-26a5-48f6-9d3b-453c1c6c2847/volumes" Nov 23 21:03:05 crc kubenswrapper[4726]: I1123 21:03:05.796537 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:05 crc kubenswrapper[4726]: I1123 21:03:05.796851 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="ceilometer-central-agent" containerID="cri-o://5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5" gracePeriod=30 Nov 23 21:03:05 crc kubenswrapper[4726]: I1123 21:03:05.797335 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="proxy-httpd" containerID="cri-o://af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f" gracePeriod=30 Nov 23 21:03:05 crc kubenswrapper[4726]: I1123 21:03:05.797398 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="sg-core" containerID="cri-o://a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d" gracePeriod=30 Nov 23 21:03:05 crc kubenswrapper[4726]: I1123 21:03:05.797442 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="ceilometer-notification-agent" containerID="cri-o://862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3" gracePeriod=30 Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.190048 4726 generic.go:334] "Generic (PLEG): container finished" podID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerID="af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f" exitCode=0 Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.190414 4726 generic.go:334] "Generic (PLEG): container finished" podID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerID="a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d" exitCode=2 Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.190433 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerDied","Data":"af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f"} Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.190457 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerDied","Data":"a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d"} Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.386801 4726 scope.go:117] "RemoveContainer" containerID="a255b22442ef792e115ddeb623ff5c601dfd3448f5c8b8eefc8b8bb74b582ec9" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.663005 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.768482 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-sb\") pod \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.768661 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-dns-svc\") pod \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.768705 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt4v5\" (UniqueName: \"kubernetes.io/projected/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-kube-api-access-bt4v5\") pod \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.768725 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-config\") pod \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.768833 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-openstack-edpm-ipam\") pod \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.768875 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-nb\") pod \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\" (UID: \"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef\") " Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.772797 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-kube-api-access-bt4v5" (OuterVolumeSpecName: "kube-api-access-bt4v5") pod "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" (UID: "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef"). InnerVolumeSpecName "kube-api-access-bt4v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.782470 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt4v5\" (UniqueName: \"kubernetes.io/projected/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-kube-api-access-bt4v5\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.824646 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" (UID: "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.838572 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" (UID: "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.843609 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" (UID: "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.854334 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-config" (OuterVolumeSpecName: "config") pod "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" (UID: "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.876619 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" (UID: "9970b11b-0b4c-49b6-868e-7ee0ad82a9ef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.884609 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.884645 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-config\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.884663 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.884682 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:06 crc kubenswrapper[4726]: I1123 21:03:06.884694 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.211427 4726 generic.go:334] "Generic (PLEG): container finished" podID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerID="5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5" exitCode=0 Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.211511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerDied","Data":"5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5"} Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.213756 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1cdd8341-a77e-4f3c-9c2e-e56125a6162c","Type":"ContainerStarted","Data":"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb"} Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.216650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" event={"ID":"9970b11b-0b4c-49b6-868e-7ee0ad82a9ef","Type":"ContainerDied","Data":"2c4e4d8ce47b445a8979af7d2537bf21cf5ac88e0223ccd60efca8483df9547c"} Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.216695 4726 scope.go:117] "RemoveContainer" containerID="0c54114e02a79af89484aae3ce55a2532c71913989ade53e32ce3596a125d907" Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.216776 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db79bb959-dxv4q" Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.319195 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5db79bb959-dxv4q"] Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.329477 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5db79bb959-dxv4q"] Nov 23 21:03:07 crc kubenswrapper[4726]: I1123 21:03:07.342734 4726 scope.go:117] "RemoveContainer" containerID="d916aca1be7cb270e111750a6edee56ec2ed05ff816e7f944f99b4e8c94e53e9" Nov 23 21:03:08 crc kubenswrapper[4726]: I1123 21:03:08.226544 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1cdd8341-a77e-4f3c-9c2e-e56125a6162c","Type":"ContainerStarted","Data":"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec"} Nov 23 21:03:08 crc kubenswrapper[4726]: I1123 21:03:08.263832 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.7714233200000002 podStartE2EDuration="16.263811574s" podCreationTimestamp="2025-11-23 21:02:52 +0000 UTC" firstStartedPulling="2025-11-23 21:02:54.016825266 +0000 UTC m=+3282.165866222" lastFinishedPulling="2025-11-23 21:03:06.50921352 +0000 UTC m=+3294.658254476" observedRunningTime="2025-11-23 21:03:08.255600034 +0000 UTC m=+3296.404641000" watchObservedRunningTime="2025-11-23 21:03:08.263811574 +0000 UTC m=+3296.412852540" Nov 23 21:03:08 crc kubenswrapper[4726]: I1123 21:03:08.599960 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" path="/var/lib/kubelet/pods/9970b11b-0b4c-49b6-868e-7ee0ad82a9ef/volumes" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.767375 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.833554 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-config-data\") pod \"6b15cde8-6249-48c9-a7f3-984adb91acf1\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.833610 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q27v5\" (UniqueName: \"kubernetes.io/projected/6b15cde8-6249-48c9-a7f3-984adb91acf1-kube-api-access-q27v5\") pod \"6b15cde8-6249-48c9-a7f3-984adb91acf1\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.833663 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-run-httpd\") pod \"6b15cde8-6249-48c9-a7f3-984adb91acf1\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.833694 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-log-httpd\") pod \"6b15cde8-6249-48c9-a7f3-984adb91acf1\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.833821 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-scripts\") pod \"6b15cde8-6249-48c9-a7f3-984adb91acf1\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.833871 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-ceilometer-tls-certs\") pod \"6b15cde8-6249-48c9-a7f3-984adb91acf1\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.833957 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-sg-core-conf-yaml\") pod \"6b15cde8-6249-48c9-a7f3-984adb91acf1\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.833981 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-combined-ca-bundle\") pod \"6b15cde8-6249-48c9-a7f3-984adb91acf1\" (UID: \"6b15cde8-6249-48c9-a7f3-984adb91acf1\") " Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.834343 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6b15cde8-6249-48c9-a7f3-984adb91acf1" (UID: "6b15cde8-6249-48c9-a7f3-984adb91acf1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.834641 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6b15cde8-6249-48c9-a7f3-984adb91acf1" (UID: "6b15cde8-6249-48c9-a7f3-984adb91acf1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.854158 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-scripts" (OuterVolumeSpecName: "scripts") pod "6b15cde8-6249-48c9-a7f3-984adb91acf1" (UID: "6b15cde8-6249-48c9-a7f3-984adb91acf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.854177 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b15cde8-6249-48c9-a7f3-984adb91acf1-kube-api-access-q27v5" (OuterVolumeSpecName: "kube-api-access-q27v5") pod "6b15cde8-6249-48c9-a7f3-984adb91acf1" (UID: "6b15cde8-6249-48c9-a7f3-984adb91acf1"). InnerVolumeSpecName "kube-api-access-q27v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.931196 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "6b15cde8-6249-48c9-a7f3-984adb91acf1" (UID: "6b15cde8-6249-48c9-a7f3-984adb91acf1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.931330 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6b15cde8-6249-48c9-a7f3-984adb91acf1" (UID: "6b15cde8-6249-48c9-a7f3-984adb91acf1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.935639 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.935673 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.935682 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.935693 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.935703 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q27v5\" (UniqueName: \"kubernetes.io/projected/6b15cde8-6249-48c9-a7f3-984adb91acf1-kube-api-access-q27v5\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.935711 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b15cde8-6249-48c9-a7f3-984adb91acf1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:09 crc kubenswrapper[4726]: I1123 21:03:09.967576 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b15cde8-6249-48c9-a7f3-984adb91acf1" (UID: "6b15cde8-6249-48c9-a7f3-984adb91acf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.026769 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-config-data" (OuterVolumeSpecName: "config-data") pod "6b15cde8-6249-48c9-a7f3-984adb91acf1" (UID: "6b15cde8-6249-48c9-a7f3-984adb91acf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.037483 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.037515 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b15cde8-6249-48c9-a7f3-984adb91acf1-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.270565 4726 generic.go:334] "Generic (PLEG): container finished" podID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerID="862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3" exitCode=0 Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.270649 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.270953 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerDied","Data":"862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3"} Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.271080 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b15cde8-6249-48c9-a7f3-984adb91acf1","Type":"ContainerDied","Data":"38ddf454ca40038cd24e949f606fce0f417db1ad06e763f43a7003e22ce5f246"} Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.271167 4726 scope.go:117] "RemoveContainer" containerID="af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.323296 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.323410 4726 scope.go:117] "RemoveContainer" containerID="a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.349949 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.365435 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.365860 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="ceilometer-central-agent" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.365884 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="ceilometer-central-agent" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.365903 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerName="horizon-log" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.365909 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerName="horizon-log" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.365922 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="proxy-httpd" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.365928 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="proxy-httpd" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.365950 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" containerName="dnsmasq-dns" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.365955 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" containerName="dnsmasq-dns" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.365968 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerName="horizon" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.365973 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerName="horizon" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.365984 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerName="horizon" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.365990 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerName="horizon" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.365998 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerName="horizon-log" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366004 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerName="horizon-log" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.366012 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="sg-core" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366017 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="sg-core" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.366033 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="ceilometer-notification-agent" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366038 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="ceilometer-notification-agent" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.366050 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" containerName="init" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366056 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" containerName="init" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366212 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="sg-core" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366221 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerName="horizon-log" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366233 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="ceilometer-central-agent" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366244 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0281d2a-1a6f-457e-b010-f52a1d8500d7" containerName="horizon" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366258 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9970b11b-0b4c-49b6-868e-7ee0ad82a9ef" containerName="dnsmasq-dns" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366268 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="proxy-httpd" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366277 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" containerName="ceilometer-notification-agent" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366289 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerName="horizon" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.366304 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="088fec86-26a5-48f6-9d3b-453c1c6c2847" containerName="horizon-log" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.367834 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.367927 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.390059 4726 scope.go:117] "RemoveContainer" containerID="862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.400823 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.401328 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.419415 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.421807 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-n2dw6 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-n2dw6 log-httpd run-httpd scripts sg-core-conf-yaml]: context canceled" pod="openstack/ceilometer-0" podUID="09e3e650-3a2a-4218-ae58-041a0c991439" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.447092 4726 scope.go:117] "RemoveContainer" containerID="5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451020 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451295 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-config-data\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451391 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451444 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-log-httpd\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451598 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2dw6\" (UniqueName: \"kubernetes.io/projected/09e3e650-3a2a-4218-ae58-041a0c991439-kube-api-access-n2dw6\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451655 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451706 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-scripts\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451750 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.451910 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-run-httpd\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.486166 4726 scope.go:117] "RemoveContainer" containerID="af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.492057 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f\": container with ID starting with af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f not found: ID does not exist" containerID="af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.492099 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f"} err="failed to get container status \"af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f\": rpc error: code = NotFound desc = could not find container \"af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f\": container with ID starting with af96dddf39556f2c4ccd6cfc098f8b510d2f3300dff3e3cdee2ca0ae7998c55f not found: ID does not exist" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.492120 4726 scope.go:117] "RemoveContainer" containerID="a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.493182 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d\": container with ID starting with a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d not found: ID does not exist" containerID="a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.493208 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d"} err="failed to get container status \"a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d\": rpc error: code = NotFound desc = could not find container \"a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d\": container with ID starting with a1df1cd5789a816c3e98e3990c9c852e90dca3653356d34511502d3d434eb18d not found: ID does not exist" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.493226 4726 scope.go:117] "RemoveContainer" containerID="862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.493615 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3\": container with ID starting with 862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3 not found: ID does not exist" containerID="862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.493633 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3"} err="failed to get container status \"862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3\": rpc error: code = NotFound desc = could not find container \"862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3\": container with ID starting with 862b5a81bc3373f775f7792ab5d2260d5e5647f88b83e3711ded19a95205d8b3 not found: ID does not exist" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.493646 4726 scope.go:117] "RemoveContainer" containerID="5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5" Nov 23 21:03:10 crc kubenswrapper[4726]: E1123 21:03:10.493811 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5\": container with ID starting with 5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5 not found: ID does not exist" containerID="5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.493827 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5"} err="failed to get container status \"5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5\": rpc error: code = NotFound desc = could not find container \"5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5\": container with ID starting with 5763430137eb423aeba9b3703c87426769087a4a528cf975157e82d7ba82b2d5 not found: ID does not exist" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.553591 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-run-httpd\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.553669 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-config-data\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.553712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.553745 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-log-httpd\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.553802 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2dw6\" (UniqueName: \"kubernetes.io/projected/09e3e650-3a2a-4218-ae58-041a0c991439-kube-api-access-n2dw6\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.553830 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.553848 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-scripts\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.553868 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.554675 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-log-httpd\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.559039 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.563102 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-scripts\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.563442 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.566727 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-run-httpd\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.570054 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.572627 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2dw6\" (UniqueName: \"kubernetes.io/projected/09e3e650-3a2a-4218-ae58-041a0c991439-kube-api-access-n2dw6\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.579993 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-config-data\") pod \"ceilometer-0\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " pod="openstack/ceilometer-0" Nov 23 21:03:10 crc kubenswrapper[4726]: I1123 21:03:10.606073 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b15cde8-6249-48c9-a7f3-984adb91acf1" path="/var/lib/kubelet/pods/6b15cde8-6249-48c9-a7f3-984adb91acf1/volumes" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.286856 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.306135 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.476306 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-config-data\") pod \"09e3e650-3a2a-4218-ae58-041a0c991439\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.476386 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-run-httpd\") pod \"09e3e650-3a2a-4218-ae58-041a0c991439\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.476404 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-log-httpd\") pod \"09e3e650-3a2a-4218-ae58-041a0c991439\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.476442 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-combined-ca-bundle\") pod \"09e3e650-3a2a-4218-ae58-041a0c991439\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.476509 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-ceilometer-tls-certs\") pod \"09e3e650-3a2a-4218-ae58-041a0c991439\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.476561 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2dw6\" (UniqueName: \"kubernetes.io/projected/09e3e650-3a2a-4218-ae58-041a0c991439-kube-api-access-n2dw6\") pod \"09e3e650-3a2a-4218-ae58-041a0c991439\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.476616 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-scripts\") pod \"09e3e650-3a2a-4218-ae58-041a0c991439\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.476671 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-sg-core-conf-yaml\") pod \"09e3e650-3a2a-4218-ae58-041a0c991439\" (UID: \"09e3e650-3a2a-4218-ae58-041a0c991439\") " Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.479434 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "09e3e650-3a2a-4218-ae58-041a0c991439" (UID: "09e3e650-3a2a-4218-ae58-041a0c991439"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.479790 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "09e3e650-3a2a-4218-ae58-041a0c991439" (UID: "09e3e650-3a2a-4218-ae58-041a0c991439"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.483693 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-config-data" (OuterVolumeSpecName: "config-data") pod "09e3e650-3a2a-4218-ae58-041a0c991439" (UID: "09e3e650-3a2a-4218-ae58-041a0c991439"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.485304 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "09e3e650-3a2a-4218-ae58-041a0c991439" (UID: "09e3e650-3a2a-4218-ae58-041a0c991439"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.486040 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "09e3e650-3a2a-4218-ae58-041a0c991439" (UID: "09e3e650-3a2a-4218-ae58-041a0c991439"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.489039 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-scripts" (OuterVolumeSpecName: "scripts") pod "09e3e650-3a2a-4218-ae58-041a0c991439" (UID: "09e3e650-3a2a-4218-ae58-041a0c991439"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.499103 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e3e650-3a2a-4218-ae58-041a0c991439-kube-api-access-n2dw6" (OuterVolumeSpecName: "kube-api-access-n2dw6") pod "09e3e650-3a2a-4218-ae58-041a0c991439" (UID: "09e3e650-3a2a-4218-ae58-041a0c991439"). InnerVolumeSpecName "kube-api-access-n2dw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.505474 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09e3e650-3a2a-4218-ae58-041a0c991439" (UID: "09e3e650-3a2a-4218-ae58-041a0c991439"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.578589 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.578802 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.578918 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.578988 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09e3e650-3a2a-4218-ae58-041a0c991439-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.579058 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.579158 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.579234 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2dw6\" (UniqueName: \"kubernetes.io/projected/09e3e650-3a2a-4218-ae58-041a0c991439-kube-api-access-n2dw6\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.579307 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09e3e650-3a2a-4218-ae58-041a0c991439-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.955803 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:03:11 crc kubenswrapper[4726]: I1123 21:03:11.972396 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.297561 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.381342 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.400555 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.414138 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.416620 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.421470 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.424407 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.424789 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.424962 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.533987 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.608141 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.608177 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3f087de-9790-4860-b3bd-1ab9e677d4bd-run-httpd\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.608205 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-scripts\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.608257 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.608294 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3f087de-9790-4860-b3bd-1ab9e677d4bd-log-httpd\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.608320 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-config-data\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.608347 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.608402 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhqxp\" (UniqueName: \"kubernetes.io/projected/a3f087de-9790-4860-b3bd-1ab9e677d4bd-kube-api-access-lhqxp\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.651575 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e3e650-3a2a-4218-ae58-041a0c991439" path="/var/lib/kubelet/pods/09e3e650-3a2a-4218-ae58-041a0c991439/volumes" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710066 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-scripts\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710145 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710191 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3f087de-9790-4860-b3bd-1ab9e677d4bd-log-httpd\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710227 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-config-data\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710261 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710322 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhqxp\" (UniqueName: \"kubernetes.io/projected/a3f087de-9790-4860-b3bd-1ab9e677d4bd-kube-api-access-lhqxp\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710398 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710416 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3f087de-9790-4860-b3bd-1ab9e677d4bd-run-httpd\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.710841 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3f087de-9790-4860-b3bd-1ab9e677d4bd-run-httpd\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.712024 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.712370 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3f087de-9790-4860-b3bd-1ab9e677d4bd-log-httpd\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.713809 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.713843 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.718917 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.724046 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-scripts\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.725463 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.726558 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-config-data\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.727542 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3f087de-9790-4860-b3bd-1ab9e677d4bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:12 crc kubenswrapper[4726]: I1123 21:03:12.758781 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhqxp\" (UniqueName: \"kubernetes.io/projected/a3f087de-9790-4860-b3bd-1ab9e677d4bd-kube-api-access-lhqxp\") pod \"ceilometer-0\" (UID: \"a3f087de-9790-4860-b3bd-1ab9e677d4bd\") " pod="openstack/ceilometer-0" Nov 23 21:03:13 crc kubenswrapper[4726]: I1123 21:03:13.037169 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 23 21:03:13 crc kubenswrapper[4726]: I1123 21:03:13.908690 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 23 21:03:14 crc kubenswrapper[4726]: I1123 21:03:14.315660 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3f087de-9790-4860-b3bd-1ab9e677d4bd","Type":"ContainerStarted","Data":"12ca8a8d2b3ff2d385dc99580e9532ac728e997080fbb7c4984114c8b5566de8"} Nov 23 21:03:14 crc kubenswrapper[4726]: I1123 21:03:14.362013 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5c65c799bb-97hbb" Nov 23 21:03:14 crc kubenswrapper[4726]: I1123 21:03:14.491610 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5c4959b8fd-6kjm6"] Nov 23 21:03:14 crc kubenswrapper[4726]: I1123 21:03:14.491870 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon-log" containerID="cri-o://be151bd7b8569f91e73ee2e0e412fce84692f7f3c3868f14f8343c70bd195ee0" gracePeriod=30 Nov 23 21:03:14 crc kubenswrapper[4726]: I1123 21:03:14.492423 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" containerID="cri-o://ce24014cd6ad57c528c327d89f0076a660767b4d3896f6e542a9195d1ce95990" gracePeriod=30 Nov 23 21:03:14 crc kubenswrapper[4726]: I1123 21:03:14.530276 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Nov 23 21:03:15 crc kubenswrapper[4726]: I1123 21:03:15.066465 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 23 21:03:15 crc kubenswrapper[4726]: I1123 21:03:15.123221 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:03:15 crc kubenswrapper[4726]: I1123 21:03:15.324700 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3f087de-9790-4860-b3bd-1ab9e677d4bd","Type":"ContainerStarted","Data":"c28ab3dfb7a429cbb06ad42b5365ba2c49206e56210604e9ce46f3d79e1377af"} Nov 23 21:03:15 crc kubenswrapper[4726]: I1123 21:03:15.324815 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerName="manila-scheduler" containerID="cri-o://fb85c536faa2e9eb2b4f98bedfc281a76086eaf19cf9a9260d7b9b84775c5779" gracePeriod=30 Nov 23 21:03:15 crc kubenswrapper[4726]: I1123 21:03:15.326151 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerName="probe" containerID="cri-o://b261223dc49063d972c9c3d098c1fc66516d9fd957e1b81a41cea2d352b250aa" gracePeriod=30 Nov 23 21:03:16 crc kubenswrapper[4726]: I1123 21:03:16.336729 4726 generic.go:334] "Generic (PLEG): container finished" podID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerID="b261223dc49063d972c9c3d098c1fc66516d9fd957e1b81a41cea2d352b250aa" exitCode=0 Nov 23 21:03:16 crc kubenswrapper[4726]: I1123 21:03:16.337239 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c3f104ff-1e92-48e0-a3c8-7ec06184f578","Type":"ContainerDied","Data":"b261223dc49063d972c9c3d098c1fc66516d9fd957e1b81a41cea2d352b250aa"} Nov 23 21:03:16 crc kubenswrapper[4726]: I1123 21:03:16.338853 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3f087de-9790-4860-b3bd-1ab9e677d4bd","Type":"ContainerStarted","Data":"d183beed79091e5aa2afba10a04a2226219f8f9095862ada2d428e0af4d8e675"} Nov 23 21:03:16 crc kubenswrapper[4726]: I1123 21:03:16.338892 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3f087de-9790-4860-b3bd-1ab9e677d4bd","Type":"ContainerStarted","Data":"ceeea1969feb60db75e3ad635eac8e686e98992e33d04fb179c445a4ed11e125"} Nov 23 21:03:17 crc kubenswrapper[4726]: I1123 21:03:17.652678 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:34082->10.217.0.243:8443: read: connection reset by peer" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.368627 4726 generic.go:334] "Generic (PLEG): container finished" podID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerID="fb85c536faa2e9eb2b4f98bedfc281a76086eaf19cf9a9260d7b9b84775c5779" exitCode=0 Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.368713 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c3f104ff-1e92-48e0-a3c8-7ec06184f578","Type":"ContainerDied","Data":"fb85c536faa2e9eb2b4f98bedfc281a76086eaf19cf9a9260d7b9b84775c5779"} Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.373848 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3f087de-9790-4860-b3bd-1ab9e677d4bd","Type":"ContainerStarted","Data":"8d0dd8b285d6e7faa2f4289ab13ebfaef7e8467445723c42863cc87a9a66d117"} Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.375298 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.376968 4726 generic.go:334] "Generic (PLEG): container finished" podID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerID="ce24014cd6ad57c528c327d89f0076a660767b4d3896f6e542a9195d1ce95990" exitCode=0 Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.377005 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c4959b8fd-6kjm6" event={"ID":"cd63a64c-bc17-40ad-abde-de910ebbe0eb","Type":"ContainerDied","Data":"ce24014cd6ad57c528c327d89f0076a660767b4d3896f6e542a9195d1ce95990"} Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.408179 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.991865974 podStartE2EDuration="6.408160239s" podCreationTimestamp="2025-11-23 21:03:12 +0000 UTC" firstStartedPulling="2025-11-23 21:03:13.918508186 +0000 UTC m=+3302.067549142" lastFinishedPulling="2025-11-23 21:03:17.334802451 +0000 UTC m=+3305.483843407" observedRunningTime="2025-11-23 21:03:18.39528978 +0000 UTC m=+3306.544330756" watchObservedRunningTime="2025-11-23 21:03:18.408160239 +0000 UTC m=+3306.557201185" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.536275 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.640409 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c3f104ff-1e92-48e0-a3c8-7ec06184f578-etc-machine-id\") pod \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.640495 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-scripts\") pod \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.640564 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f104ff-1e92-48e0-a3c8-7ec06184f578-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c3f104ff-1e92-48e0-a3c8-7ec06184f578" (UID: "c3f104ff-1e92-48e0-a3c8-7ec06184f578"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.640677 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-combined-ca-bundle\") pod \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.640726 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data\") pod \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.640979 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data-custom\") pod \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.641022 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p667h\" (UniqueName: \"kubernetes.io/projected/c3f104ff-1e92-48e0-a3c8-7ec06184f578-kube-api-access-p667h\") pod \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\" (UID: \"c3f104ff-1e92-48e0-a3c8-7ec06184f578\") " Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.641668 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c3f104ff-1e92-48e0-a3c8-7ec06184f578-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.654843 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-scripts" (OuterVolumeSpecName: "scripts") pod "c3f104ff-1e92-48e0-a3c8-7ec06184f578" (UID: "c3f104ff-1e92-48e0-a3c8-7ec06184f578"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.655101 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3f104ff-1e92-48e0-a3c8-7ec06184f578-kube-api-access-p667h" (OuterVolumeSpecName: "kube-api-access-p667h") pod "c3f104ff-1e92-48e0-a3c8-7ec06184f578" (UID: "c3f104ff-1e92-48e0-a3c8-7ec06184f578"). InnerVolumeSpecName "kube-api-access-p667h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.655257 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c3f104ff-1e92-48e0-a3c8-7ec06184f578" (UID: "c3f104ff-1e92-48e0-a3c8-7ec06184f578"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.718410 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3f104ff-1e92-48e0-a3c8-7ec06184f578" (UID: "c3f104ff-1e92-48e0-a3c8-7ec06184f578"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.743007 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.743335 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.743477 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.743577 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p667h\" (UniqueName: \"kubernetes.io/projected/c3f104ff-1e92-48e0-a3c8-7ec06184f578-kube-api-access-p667h\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.759017 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data" (OuterVolumeSpecName: "config-data") pod "c3f104ff-1e92-48e0-a3c8-7ec06184f578" (UID: "c3f104ff-1e92-48e0-a3c8-7ec06184f578"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:18 crc kubenswrapper[4726]: I1123 21:03:18.845090 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3f104ff-1e92-48e0-a3c8-7ec06184f578-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.082224 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.243:8443: connect: connection refused" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.389955 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"c3f104ff-1e92-48e0-a3c8-7ec06184f578","Type":"ContainerDied","Data":"bafe90b33ca8f8c6b265879c0290bc56d0900013b580fa7788f47d1d209bc385"} Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.390340 4726 scope.go:117] "RemoveContainer" containerID="b261223dc49063d972c9c3d098c1fc66516d9fd957e1b81a41cea2d352b250aa" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.390917 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.412032 4726 scope.go:117] "RemoveContainer" containerID="fb85c536faa2e9eb2b4f98bedfc281a76086eaf19cf9a9260d7b9b84775c5779" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.454538 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.482111 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.495592 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:03:19 crc kubenswrapper[4726]: E1123 21:03:19.496798 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerName="manila-scheduler" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.496842 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerName="manila-scheduler" Nov 23 21:03:19 crc kubenswrapper[4726]: E1123 21:03:19.496979 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerName="probe" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.496990 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerName="probe" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.498024 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerName="manila-scheduler" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.498062 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" containerName="probe" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.500237 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.508529 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.531192 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.662958 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.663012 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-config-data\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.663039 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ba73f5-3651-4b4b-9f8c-58fb451a655f-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.663076 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.663119 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-scripts\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.663154 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzdcm\" (UniqueName: \"kubernetes.io/projected/78ba73f5-3651-4b4b-9f8c-58fb451a655f-kube-api-access-wzdcm\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.765233 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-scripts\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.765318 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzdcm\" (UniqueName: \"kubernetes.io/projected/78ba73f5-3651-4b4b-9f8c-58fb451a655f-kube-api-access-wzdcm\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.765411 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.765441 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-config-data\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.765458 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ba73f5-3651-4b4b-9f8c-58fb451a655f-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.765493 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.767165 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ba73f5-3651-4b4b-9f8c-58fb451a655f-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.772396 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-scripts\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.780904 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-config-data\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.782289 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.782450 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78ba73f5-3651-4b4b-9f8c-58fb451a655f-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.788824 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzdcm\" (UniqueName: \"kubernetes.io/projected/78ba73f5-3651-4b4b-9f8c-58fb451a655f-kube-api-access-wzdcm\") pod \"manila-scheduler-0\" (UID: \"78ba73f5-3651-4b4b-9f8c-58fb451a655f\") " pod="openstack/manila-scheduler-0" Nov 23 21:03:19 crc kubenswrapper[4726]: I1123 21:03:19.843383 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 23 21:03:20 crc kubenswrapper[4726]: I1123 21:03:20.289518 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 23 21:03:20 crc kubenswrapper[4726]: I1123 21:03:20.400923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"78ba73f5-3651-4b4b-9f8c-58fb451a655f","Type":"ContainerStarted","Data":"e4cb8425f3f0c132533eed517dd1731c455b89662e1f1f737cc0da9c8a02a640"} Nov 23 21:03:20 crc kubenswrapper[4726]: I1123 21:03:20.604410 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3f104ff-1e92-48e0-a3c8-7ec06184f578" path="/var/lib/kubelet/pods/c3f104ff-1e92-48e0-a3c8-7ec06184f578/volumes" Nov 23 21:03:21 crc kubenswrapper[4726]: I1123 21:03:21.410100 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"78ba73f5-3651-4b4b-9f8c-58fb451a655f","Type":"ContainerStarted","Data":"3bc1f70bbefd94ef0fbf246111e7a5a5a1f5df6400cf9ddadf23df8697214cdd"} Nov 23 21:03:21 crc kubenswrapper[4726]: I1123 21:03:21.410478 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"78ba73f5-3651-4b4b-9f8c-58fb451a655f","Type":"ContainerStarted","Data":"ef8ae6b39e236bc231757ae85b90b724c7686a95e4d803c415570978d69e63d5"} Nov 23 21:03:21 crc kubenswrapper[4726]: I1123 21:03:21.436635 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.43661822 podStartE2EDuration="2.43661822s" podCreationTimestamp="2025-11-23 21:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:03:21.434164201 +0000 UTC m=+3309.583205157" watchObservedRunningTime="2025-11-23 21:03:21.43661822 +0000 UTC m=+3309.585659176" Nov 23 21:03:21 crc kubenswrapper[4726]: I1123 21:03:21.996288 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 23 21:03:24 crc kubenswrapper[4726]: I1123 21:03:24.068434 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 23 21:03:24 crc kubenswrapper[4726]: I1123 21:03:24.142081 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:03:24 crc kubenswrapper[4726]: I1123 21:03:24.435072 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerName="manila-share" containerID="cri-o://2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb" gracePeriod=30 Nov 23 21:03:24 crc kubenswrapper[4726]: I1123 21:03:24.435100 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerName="probe" containerID="cri-o://579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec" gracePeriod=30 Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.372446 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.443754 4726 generic.go:334] "Generic (PLEG): container finished" podID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerID="579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec" exitCode=0 Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.443792 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.443799 4726 generic.go:334] "Generic (PLEG): container finished" podID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerID="2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb" exitCode=1 Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.443821 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1cdd8341-a77e-4f3c-9c2e-e56125a6162c","Type":"ContainerDied","Data":"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec"} Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.443846 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1cdd8341-a77e-4f3c-9c2e-e56125a6162c","Type":"ContainerDied","Data":"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb"} Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.443858 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1cdd8341-a77e-4f3c-9c2e-e56125a6162c","Type":"ContainerDied","Data":"c2be0777839febc93886ece1b47a147fee2b9dfab9a4ffbdab002f5c58fb767a"} Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.443891 4726 scope.go:117] "RemoveContainer" containerID="579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.473827 4726 scope.go:117] "RemoveContainer" containerID="2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.490159 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-etc-machine-id\") pod \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.490207 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4q8c\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-kube-api-access-b4q8c\") pod \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.490248 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-var-lib-manila\") pod \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.490299 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-combined-ca-bundle\") pod \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.490371 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-ceph\") pod \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.490415 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-scripts\") pod \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.490448 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data\") pod \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.490547 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data-custom\") pod \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\" (UID: \"1cdd8341-a77e-4f3c-9c2e-e56125a6162c\") " Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.491249 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "1cdd8341-a77e-4f3c-9c2e-e56125a6162c" (UID: "1cdd8341-a77e-4f3c-9c2e-e56125a6162c"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.491950 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1cdd8341-a77e-4f3c-9c2e-e56125a6162c" (UID: "1cdd8341-a77e-4f3c-9c2e-e56125a6162c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.496888 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-scripts" (OuterVolumeSpecName: "scripts") pod "1cdd8341-a77e-4f3c-9c2e-e56125a6162c" (UID: "1cdd8341-a77e-4f3c-9c2e-e56125a6162c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.504045 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-ceph" (OuterVolumeSpecName: "ceph") pod "1cdd8341-a77e-4f3c-9c2e-e56125a6162c" (UID: "1cdd8341-a77e-4f3c-9c2e-e56125a6162c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.524609 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1cdd8341-a77e-4f3c-9c2e-e56125a6162c" (UID: "1cdd8341-a77e-4f3c-9c2e-e56125a6162c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.527652 4726 scope.go:117] "RemoveContainer" containerID="579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec" Nov 23 21:03:25 crc kubenswrapper[4726]: E1123 21:03:25.531707 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec\": container with ID starting with 579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec not found: ID does not exist" containerID="579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.531837 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec"} err="failed to get container status \"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec\": rpc error: code = NotFound desc = could not find container \"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec\": container with ID starting with 579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec not found: ID does not exist" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.533025 4726 scope.go:117] "RemoveContainer" containerID="2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.533685 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-kube-api-access-b4q8c" (OuterVolumeSpecName: "kube-api-access-b4q8c") pod "1cdd8341-a77e-4f3c-9c2e-e56125a6162c" (UID: "1cdd8341-a77e-4f3c-9c2e-e56125a6162c"). InnerVolumeSpecName "kube-api-access-b4q8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:25 crc kubenswrapper[4726]: E1123 21:03:25.541119 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb\": container with ID starting with 2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb not found: ID does not exist" containerID="2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.541172 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb"} err="failed to get container status \"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb\": rpc error: code = NotFound desc = could not find container \"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb\": container with ID starting with 2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb not found: ID does not exist" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.541208 4726 scope.go:117] "RemoveContainer" containerID="579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.541699 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec"} err="failed to get container status \"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec\": rpc error: code = NotFound desc = could not find container \"579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec\": container with ID starting with 579cd7f56a03e3b9e35a709042b2f46bf51d786703684f7243e3e34d904038ec not found: ID does not exist" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.541732 4726 scope.go:117] "RemoveContainer" containerID="2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.541974 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb"} err="failed to get container status \"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb\": rpc error: code = NotFound desc = could not find container \"2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb\": container with ID starting with 2b5b4cb5122dc5cdda77544a26585cc282853a36b9387a0cb703d8ecfc5da2bb not found: ID does not exist" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.578190 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1cdd8341-a77e-4f3c-9c2e-e56125a6162c" (UID: "1cdd8341-a77e-4f3c-9c2e-e56125a6162c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.592427 4726 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-ceph\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.592465 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.592478 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.592491 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.592503 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4q8c\" (UniqueName: \"kubernetes.io/projected/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-kube-api-access-b4q8c\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.592517 4726 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-var-lib-manila\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.592529 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.628436 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data" (OuterVolumeSpecName: "config-data") pod "1cdd8341-a77e-4f3c-9c2e-e56125a6162c" (UID: "1cdd8341-a77e-4f3c-9c2e-e56125a6162c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.694604 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cdd8341-a77e-4f3c-9c2e-e56125a6162c-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.820300 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.831257 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.852289 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:03:25 crc kubenswrapper[4726]: E1123 21:03:25.852768 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerName="manila-share" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.852827 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerName="manila-share" Nov 23 21:03:25 crc kubenswrapper[4726]: E1123 21:03:25.852855 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerName="probe" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.852866 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerName="probe" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.853095 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerName="manila-share" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.853121 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" containerName="probe" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.854340 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.858067 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.905976 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.906035 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-scripts\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.906061 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.906097 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-ceph\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.906119 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.906144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-config-data\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.906168 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksghw\" (UniqueName: \"kubernetes.io/projected/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-kube-api-access-ksghw\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.906193 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:25 crc kubenswrapper[4726]: I1123 21:03:25.909863 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.007572 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-ceph\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.007614 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.007635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-config-data\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.007655 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksghw\" (UniqueName: \"kubernetes.io/projected/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-kube-api-access-ksghw\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.007677 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.007788 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.007806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.007820 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-scripts\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.008428 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.008733 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.011697 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-scripts\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.011777 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.012985 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-ceph\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.014537 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.017526 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-config-data\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.028165 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksghw\" (UniqueName: \"kubernetes.io/projected/ccbab0a3-e42f-42fd-8adb-ed1bae00642c-kube-api-access-ksghw\") pod \"manila-share-share1-0\" (UID: \"ccbab0a3-e42f-42fd-8adb-ed1bae00642c\") " pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.182662 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.598742 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cdd8341-a77e-4f3c-9c2e-e56125a6162c" path="/var/lib/kubelet/pods/1cdd8341-a77e-4f3c-9c2e-e56125a6162c/volumes" Nov 23 21:03:26 crc kubenswrapper[4726]: I1123 21:03:26.750579 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 23 21:03:26 crc kubenswrapper[4726]: W1123 21:03:26.766100 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccbab0a3_e42f_42fd_8adb_ed1bae00642c.slice/crio-17bd90d70913c126172a091d7baead31e01f92818a1adce00b77ebb7f19930d9 WatchSource:0}: Error finding container 17bd90d70913c126172a091d7baead31e01f92818a1adce00b77ebb7f19930d9: Status 404 returned error can't find the container with id 17bd90d70913c126172a091d7baead31e01f92818a1adce00b77ebb7f19930d9 Nov 23 21:03:27 crc kubenswrapper[4726]: I1123 21:03:27.489613 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ccbab0a3-e42f-42fd-8adb-ed1bae00642c","Type":"ContainerStarted","Data":"e3c2486f0213a502728c7e4274c0ff69f85a2526f27ca887fe7abc970f76e7e8"} Nov 23 21:03:27 crc kubenswrapper[4726]: I1123 21:03:27.489932 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ccbab0a3-e42f-42fd-8adb-ed1bae00642c","Type":"ContainerStarted","Data":"17bd90d70913c126172a091d7baead31e01f92818a1adce00b77ebb7f19930d9"} Nov 23 21:03:28 crc kubenswrapper[4726]: I1123 21:03:28.499842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"ccbab0a3-e42f-42fd-8adb-ed1bae00642c","Type":"ContainerStarted","Data":"28401f997dcf5a37d33ca69fea59d6683f3bd20552e599dcad92eebb18300a81"} Nov 23 21:03:28 crc kubenswrapper[4726]: I1123 21:03:28.519792 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.519776743 podStartE2EDuration="3.519776743s" podCreationTimestamp="2025-11-23 21:03:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:03:28.517114378 +0000 UTC m=+3316.666155344" watchObservedRunningTime="2025-11-23 21:03:28.519776743 +0000 UTC m=+3316.668817699" Nov 23 21:03:29 crc kubenswrapper[4726]: I1123 21:03:29.082306 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.243:8443: connect: connection refused" Nov 23 21:03:29 crc kubenswrapper[4726]: I1123 21:03:29.844889 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 23 21:03:36 crc kubenswrapper[4726]: I1123 21:03:36.183278 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 23 21:03:39 crc kubenswrapper[4726]: I1123 21:03:39.043457 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:03:39 crc kubenswrapper[4726]: I1123 21:03:39.044176 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:03:39 crc kubenswrapper[4726]: I1123 21:03:39.082589 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5c4959b8fd-6kjm6" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.243:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.243:8443: connect: connection refused" Nov 23 21:03:41 crc kubenswrapper[4726]: I1123 21:03:41.393742 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 23 21:03:43 crc kubenswrapper[4726]: I1123 21:03:43.053039 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 23 21:03:44 crc kubenswrapper[4726]: I1123 21:03:44.704041 4726 generic.go:334] "Generic (PLEG): container finished" podID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerID="be151bd7b8569f91e73ee2e0e412fce84692f7f3c3868f14f8343c70bd195ee0" exitCode=137 Nov 23 21:03:44 crc kubenswrapper[4726]: I1123 21:03:44.704483 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c4959b8fd-6kjm6" event={"ID":"cd63a64c-bc17-40ad-abde-de910ebbe0eb","Type":"ContainerDied","Data":"be151bd7b8569f91e73ee2e0e412fce84692f7f3c3868f14f8343c70bd195ee0"} Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.024442 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.196025 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-tls-certs\") pod \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.196375 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-secret-key\") pod \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.196424 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-config-data\") pod \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.196547 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-combined-ca-bundle\") pod \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.196570 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-scripts\") pod \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.196595 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd63a64c-bc17-40ad-abde-de910ebbe0eb-logs\") pod \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.196696 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5q2z\" (UniqueName: \"kubernetes.io/projected/cd63a64c-bc17-40ad-abde-de910ebbe0eb-kube-api-access-x5q2z\") pod \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\" (UID: \"cd63a64c-bc17-40ad-abde-de910ebbe0eb\") " Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.197814 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd63a64c-bc17-40ad-abde-de910ebbe0eb-logs" (OuterVolumeSpecName: "logs") pod "cd63a64c-bc17-40ad-abde-de910ebbe0eb" (UID: "cd63a64c-bc17-40ad-abde-de910ebbe0eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.203246 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "cd63a64c-bc17-40ad-abde-de910ebbe0eb" (UID: "cd63a64c-bc17-40ad-abde-de910ebbe0eb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.203280 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd63a64c-bc17-40ad-abde-de910ebbe0eb-kube-api-access-x5q2z" (OuterVolumeSpecName: "kube-api-access-x5q2z") pod "cd63a64c-bc17-40ad-abde-de910ebbe0eb" (UID: "cd63a64c-bc17-40ad-abde-de910ebbe0eb"). InnerVolumeSpecName "kube-api-access-x5q2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.225705 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-config-data" (OuterVolumeSpecName: "config-data") pod "cd63a64c-bc17-40ad-abde-de910ebbe0eb" (UID: "cd63a64c-bc17-40ad-abde-de910ebbe0eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.226610 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd63a64c-bc17-40ad-abde-de910ebbe0eb" (UID: "cd63a64c-bc17-40ad-abde-de910ebbe0eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.233614 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-scripts" (OuterVolumeSpecName: "scripts") pod "cd63a64c-bc17-40ad-abde-de910ebbe0eb" (UID: "cd63a64c-bc17-40ad-abde-de910ebbe0eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.273368 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "cd63a64c-bc17-40ad-abde-de910ebbe0eb" (UID: "cd63a64c-bc17-40ad-abde-de910ebbe0eb"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.299131 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.299323 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-scripts\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.299378 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd63a64c-bc17-40ad-abde-de910ebbe0eb-logs\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.299430 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5q2z\" (UniqueName: \"kubernetes.io/projected/cd63a64c-bc17-40ad-abde-de910ebbe0eb-kube-api-access-x5q2z\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.299491 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.299542 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd63a64c-bc17-40ad-abde-de910ebbe0eb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.299592 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd63a64c-bc17-40ad-abde-de910ebbe0eb-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.733055 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c4959b8fd-6kjm6" event={"ID":"cd63a64c-bc17-40ad-abde-de910ebbe0eb","Type":"ContainerDied","Data":"3b857eb0eb7ac38e02fd46410a19c2ed479a5d5890a582a8441f1100b4156847"} Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.733141 4726 scope.go:117] "RemoveContainer" containerID="ce24014cd6ad57c528c327d89f0076a660767b4d3896f6e542a9195d1ce95990" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.733093 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c4959b8fd-6kjm6" Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.810064 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5c4959b8fd-6kjm6"] Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.817564 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5c4959b8fd-6kjm6"] Nov 23 21:03:45 crc kubenswrapper[4726]: I1123 21:03:45.986842 4726 scope.go:117] "RemoveContainer" containerID="be151bd7b8569f91e73ee2e0e412fce84692f7f3c3868f14f8343c70bd195ee0" Nov 23 21:03:46 crc kubenswrapper[4726]: I1123 21:03:46.600475 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" path="/var/lib/kubelet/pods/cd63a64c-bc17-40ad-abde-de910ebbe0eb/volumes" Nov 23 21:03:47 crc kubenswrapper[4726]: I1123 21:03:47.661999 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.159584 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rgxg4"] Nov 23 21:03:53 crc kubenswrapper[4726]: E1123 21:03:53.160755 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon-log" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.160776 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon-log" Nov 23 21:03:53 crc kubenswrapper[4726]: E1123 21:03:53.160816 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.160823 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.161065 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.161080 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd63a64c-bc17-40ad-abde-de910ebbe0eb" containerName="horizon-log" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.162943 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.186805 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rgxg4"] Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.268898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-catalog-content\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.269435 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-utilities\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.269539 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x45ht\" (UniqueName: \"kubernetes.io/projected/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-kube-api-access-x45ht\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.371030 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-catalog-content\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.371076 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-utilities\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.371109 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x45ht\" (UniqueName: \"kubernetes.io/projected/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-kube-api-access-x45ht\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.371663 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-catalog-content\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.371759 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-utilities\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.404609 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x45ht\" (UniqueName: \"kubernetes.io/projected/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-kube-api-access-x45ht\") pod \"redhat-operators-rgxg4\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.502154 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:03:53 crc kubenswrapper[4726]: I1123 21:03:53.975651 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rgxg4"] Nov 23 21:03:53 crc kubenswrapper[4726]: W1123 21:03:53.981968 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabb4b648_4404_4a69_a6b5_9f8c7dd0973c.slice/crio-d503e9ac96b507813beaa50af43f74b7ba154b3422bdad81ba7d6c4527a5231e WatchSource:0}: Error finding container d503e9ac96b507813beaa50af43f74b7ba154b3422bdad81ba7d6c4527a5231e: Status 404 returned error can't find the container with id d503e9ac96b507813beaa50af43f74b7ba154b3422bdad81ba7d6c4527a5231e Nov 23 21:03:54 crc kubenswrapper[4726]: I1123 21:03:54.825124 4726 generic.go:334] "Generic (PLEG): container finished" podID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerID="57a9b470a2d1a5d5b6e75cc08197b2f543fab7feec2dda2fff064dac46cb19b9" exitCode=0 Nov 23 21:03:54 crc kubenswrapper[4726]: I1123 21:03:54.825244 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgxg4" event={"ID":"abb4b648-4404-4a69-a6b5-9f8c7dd0973c","Type":"ContainerDied","Data":"57a9b470a2d1a5d5b6e75cc08197b2f543fab7feec2dda2fff064dac46cb19b9"} Nov 23 21:03:54 crc kubenswrapper[4726]: I1123 21:03:54.825586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgxg4" event={"ID":"abb4b648-4404-4a69-a6b5-9f8c7dd0973c","Type":"ContainerStarted","Data":"d503e9ac96b507813beaa50af43f74b7ba154b3422bdad81ba7d6c4527a5231e"} Nov 23 21:03:55 crc kubenswrapper[4726]: I1123 21:03:55.876030 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgxg4" event={"ID":"abb4b648-4404-4a69-a6b5-9f8c7dd0973c","Type":"ContainerStarted","Data":"333e70493fa7c2ebf866840637276e50ddfdd8dae463df53e3e24c6020163eb6"} Nov 23 21:04:00 crc kubenswrapper[4726]: I1123 21:04:00.919522 4726 generic.go:334] "Generic (PLEG): container finished" podID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerID="333e70493fa7c2ebf866840637276e50ddfdd8dae463df53e3e24c6020163eb6" exitCode=0 Nov 23 21:04:00 crc kubenswrapper[4726]: I1123 21:04:00.919615 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgxg4" event={"ID":"abb4b648-4404-4a69-a6b5-9f8c7dd0973c","Type":"ContainerDied","Data":"333e70493fa7c2ebf866840637276e50ddfdd8dae463df53e3e24c6020163eb6"} Nov 23 21:04:01 crc kubenswrapper[4726]: I1123 21:04:01.936375 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgxg4" event={"ID":"abb4b648-4404-4a69-a6b5-9f8c7dd0973c","Type":"ContainerStarted","Data":"ddcbef429e6ce9b56b45e2a2d0f9f769a69d665affe6d37cd213ff6d703493dd"} Nov 23 21:04:01 crc kubenswrapper[4726]: I1123 21:04:01.959249 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rgxg4" podStartSLOduration=2.252952723 podStartE2EDuration="8.959234707s" podCreationTimestamp="2025-11-23 21:03:53 +0000 UTC" firstStartedPulling="2025-11-23 21:03:54.827912659 +0000 UTC m=+3342.976953615" lastFinishedPulling="2025-11-23 21:04:01.534194643 +0000 UTC m=+3349.683235599" observedRunningTime="2025-11-23 21:04:01.957466638 +0000 UTC m=+3350.106507624" watchObservedRunningTime="2025-11-23 21:04:01.959234707 +0000 UTC m=+3350.108275663" Nov 23 21:04:03 crc kubenswrapper[4726]: I1123 21:04:03.502511 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:04:03 crc kubenswrapper[4726]: I1123 21:04:03.505033 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:04:04 crc kubenswrapper[4726]: I1123 21:04:04.553673 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rgxg4" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="registry-server" probeResult="failure" output=< Nov 23 21:04:04 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 21:04:04 crc kubenswrapper[4726]: > Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.043179 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.043782 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.522123 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t7hfp"] Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.525673 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.540759 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7hfp"] Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.642518 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-catalog-content\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.642898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-utilities\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.642971 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7fwb\" (UniqueName: \"kubernetes.io/projected/3bccb5da-c358-49f7-bc96-478badba2d64-kube-api-access-l7fwb\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.745467 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-utilities\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.745586 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7fwb\" (UniqueName: \"kubernetes.io/projected/3bccb5da-c358-49f7-bc96-478badba2d64-kube-api-access-l7fwb\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.745743 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-catalog-content\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.746452 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-catalog-content\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.746464 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-utilities\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.772312 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7fwb\" (UniqueName: \"kubernetes.io/projected/3bccb5da-c358-49f7-bc96-478badba2d64-kube-api-access-l7fwb\") pod \"redhat-marketplace-t7hfp\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:09 crc kubenswrapper[4726]: I1123 21:04:09.855851 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:10 crc kubenswrapper[4726]: I1123 21:04:10.310892 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7hfp"] Nov 23 21:04:11 crc kubenswrapper[4726]: I1123 21:04:11.019378 4726 generic.go:334] "Generic (PLEG): container finished" podID="3bccb5da-c358-49f7-bc96-478badba2d64" containerID="0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1" exitCode=0 Nov 23 21:04:11 crc kubenswrapper[4726]: I1123 21:04:11.019686 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7hfp" event={"ID":"3bccb5da-c358-49f7-bc96-478badba2d64","Type":"ContainerDied","Data":"0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1"} Nov 23 21:04:11 crc kubenswrapper[4726]: I1123 21:04:11.019711 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7hfp" event={"ID":"3bccb5da-c358-49f7-bc96-478badba2d64","Type":"ContainerStarted","Data":"bec3c9c58048b77621fdbfb51a67dba1bfa17f8d3ec7bb8710234262c2dbabc4"} Nov 23 21:04:12 crc kubenswrapper[4726]: I1123 21:04:12.029465 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7hfp" event={"ID":"3bccb5da-c358-49f7-bc96-478badba2d64","Type":"ContainerStarted","Data":"3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a"} Nov 23 21:04:13 crc kubenswrapper[4726]: I1123 21:04:13.057296 4726 generic.go:334] "Generic (PLEG): container finished" podID="3bccb5da-c358-49f7-bc96-478badba2d64" containerID="3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a" exitCode=0 Nov 23 21:04:13 crc kubenswrapper[4726]: I1123 21:04:13.057417 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7hfp" event={"ID":"3bccb5da-c358-49f7-bc96-478badba2d64","Type":"ContainerDied","Data":"3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a"} Nov 23 21:04:13 crc kubenswrapper[4726]: I1123 21:04:13.578187 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:04:13 crc kubenswrapper[4726]: I1123 21:04:13.674021 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:04:14 crc kubenswrapper[4726]: I1123 21:04:14.069043 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7hfp" event={"ID":"3bccb5da-c358-49f7-bc96-478badba2d64","Type":"ContainerStarted","Data":"269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b"} Nov 23 21:04:14 crc kubenswrapper[4726]: I1123 21:04:14.099078 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t7hfp" podStartSLOduration=2.6166470029999997 podStartE2EDuration="5.099056211s" podCreationTimestamp="2025-11-23 21:04:09 +0000 UTC" firstStartedPulling="2025-11-23 21:04:11.021244114 +0000 UTC m=+3359.170285070" lastFinishedPulling="2025-11-23 21:04:13.503653312 +0000 UTC m=+3361.652694278" observedRunningTime="2025-11-23 21:04:14.090406889 +0000 UTC m=+3362.239447865" watchObservedRunningTime="2025-11-23 21:04:14.099056211 +0000 UTC m=+3362.248097187" Nov 23 21:04:15 crc kubenswrapper[4726]: E1123 21:04:15.828471 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-conmon-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache]" Nov 23 21:04:15 crc kubenswrapper[4726]: I1123 21:04:15.898670 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rgxg4"] Nov 23 21:04:15 crc kubenswrapper[4726]: I1123 21:04:15.898912 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rgxg4" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="registry-server" containerID="cri-o://ddcbef429e6ce9b56b45e2a2d0f9f769a69d665affe6d37cd213ff6d703493dd" gracePeriod=2 Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.090816 4726 generic.go:334] "Generic (PLEG): container finished" podID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerID="ddcbef429e6ce9b56b45e2a2d0f9f769a69d665affe6d37cd213ff6d703493dd" exitCode=0 Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.090882 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgxg4" event={"ID":"abb4b648-4404-4a69-a6b5-9f8c7dd0973c","Type":"ContainerDied","Data":"ddcbef429e6ce9b56b45e2a2d0f9f769a69d665affe6d37cd213ff6d703493dd"} Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.395924 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.496509 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-utilities\") pod \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.496835 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-catalog-content\") pod \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.496888 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x45ht\" (UniqueName: \"kubernetes.io/projected/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-kube-api-access-x45ht\") pod \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\" (UID: \"abb4b648-4404-4a69-a6b5-9f8c7dd0973c\") " Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.497564 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-utilities" (OuterVolumeSpecName: "utilities") pod "abb4b648-4404-4a69-a6b5-9f8c7dd0973c" (UID: "abb4b648-4404-4a69-a6b5-9f8c7dd0973c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.501640 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-kube-api-access-x45ht" (OuterVolumeSpecName: "kube-api-access-x45ht") pod "abb4b648-4404-4a69-a6b5-9f8c7dd0973c" (UID: "abb4b648-4404-4a69-a6b5-9f8c7dd0973c"). InnerVolumeSpecName "kube-api-access-x45ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.582707 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abb4b648-4404-4a69-a6b5-9f8c7dd0973c" (UID: "abb4b648-4404-4a69-a6b5-9f8c7dd0973c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.599076 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.599123 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x45ht\" (UniqueName: \"kubernetes.io/projected/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-kube-api-access-x45ht\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:16 crc kubenswrapper[4726]: I1123 21:04:16.599137 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abb4b648-4404-4a69-a6b5-9f8c7dd0973c-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:17 crc kubenswrapper[4726]: I1123 21:04:17.108805 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgxg4" event={"ID":"abb4b648-4404-4a69-a6b5-9f8c7dd0973c","Type":"ContainerDied","Data":"d503e9ac96b507813beaa50af43f74b7ba154b3422bdad81ba7d6c4527a5231e"} Nov 23 21:04:17 crc kubenswrapper[4726]: I1123 21:04:17.108855 4726 scope.go:117] "RemoveContainer" containerID="ddcbef429e6ce9b56b45e2a2d0f9f769a69d665affe6d37cd213ff6d703493dd" Nov 23 21:04:17 crc kubenswrapper[4726]: I1123 21:04:17.108887 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgxg4" Nov 23 21:04:17 crc kubenswrapper[4726]: I1123 21:04:17.140565 4726 scope.go:117] "RemoveContainer" containerID="333e70493fa7c2ebf866840637276e50ddfdd8dae463df53e3e24c6020163eb6" Nov 23 21:04:17 crc kubenswrapper[4726]: I1123 21:04:17.140780 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rgxg4"] Nov 23 21:04:17 crc kubenswrapper[4726]: I1123 21:04:17.149609 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rgxg4"] Nov 23 21:04:17 crc kubenswrapper[4726]: I1123 21:04:17.163771 4726 scope.go:117] "RemoveContainer" containerID="57a9b470a2d1a5d5b6e75cc08197b2f543fab7feec2dda2fff064dac46cb19b9" Nov 23 21:04:18 crc kubenswrapper[4726]: I1123 21:04:18.608925 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" path="/var/lib/kubelet/pods/abb4b648-4404-4a69-a6b5-9f8c7dd0973c/volumes" Nov 23 21:04:19 crc kubenswrapper[4726]: I1123 21:04:19.856531 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:19 crc kubenswrapper[4726]: I1123 21:04:19.856943 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:19 crc kubenswrapper[4726]: I1123 21:04:19.915708 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:20 crc kubenswrapper[4726]: I1123 21:04:20.197984 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:21 crc kubenswrapper[4726]: I1123 21:04:21.101448 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7hfp"] Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.158800 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t7hfp" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" containerName="registry-server" containerID="cri-o://269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b" gracePeriod=2 Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.625570 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.765502 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-catalog-content\") pod \"3bccb5da-c358-49f7-bc96-478badba2d64\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.765570 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7fwb\" (UniqueName: \"kubernetes.io/projected/3bccb5da-c358-49f7-bc96-478badba2d64-kube-api-access-l7fwb\") pod \"3bccb5da-c358-49f7-bc96-478badba2d64\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.765754 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-utilities\") pod \"3bccb5da-c358-49f7-bc96-478badba2d64\" (UID: \"3bccb5da-c358-49f7-bc96-478badba2d64\") " Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.768597 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-utilities" (OuterVolumeSpecName: "utilities") pod "3bccb5da-c358-49f7-bc96-478badba2d64" (UID: "3bccb5da-c358-49f7-bc96-478badba2d64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.771795 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bccb5da-c358-49f7-bc96-478badba2d64-kube-api-access-l7fwb" (OuterVolumeSpecName: "kube-api-access-l7fwb") pod "3bccb5da-c358-49f7-bc96-478badba2d64" (UID: "3bccb5da-c358-49f7-bc96-478badba2d64"). InnerVolumeSpecName "kube-api-access-l7fwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.791028 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bccb5da-c358-49f7-bc96-478badba2d64" (UID: "3bccb5da-c358-49f7-bc96-478badba2d64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.870631 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.870831 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bccb5da-c358-49f7-bc96-478badba2d64-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:22 crc kubenswrapper[4726]: I1123 21:04:22.870854 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7fwb\" (UniqueName: \"kubernetes.io/projected/3bccb5da-c358-49f7-bc96-478badba2d64-kube-api-access-l7fwb\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.168969 4726 generic.go:334] "Generic (PLEG): container finished" podID="3bccb5da-c358-49f7-bc96-478badba2d64" containerID="269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b" exitCode=0 Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.169066 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7hfp" event={"ID":"3bccb5da-c358-49f7-bc96-478badba2d64","Type":"ContainerDied","Data":"269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b"} Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.170114 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7hfp" event={"ID":"3bccb5da-c358-49f7-bc96-478badba2d64","Type":"ContainerDied","Data":"bec3c9c58048b77621fdbfb51a67dba1bfa17f8d3ec7bb8710234262c2dbabc4"} Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.170182 4726 scope.go:117] "RemoveContainer" containerID="269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.169124 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7hfp" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.204665 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7hfp"] Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.214496 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7hfp"] Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.219264 4726 scope.go:117] "RemoveContainer" containerID="3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.250748 4726 scope.go:117] "RemoveContainer" containerID="0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.319854 4726 scope.go:117] "RemoveContainer" containerID="269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b" Nov 23 21:04:23 crc kubenswrapper[4726]: E1123 21:04:23.320503 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b\": container with ID starting with 269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b not found: ID does not exist" containerID="269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.320598 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b"} err="failed to get container status \"269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b\": rpc error: code = NotFound desc = could not find container \"269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b\": container with ID starting with 269c326acf09cd3dddfc791fa805891302bf8e0fe35af5b9fbf2f8f828d35a0b not found: ID does not exist" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.320684 4726 scope.go:117] "RemoveContainer" containerID="3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a" Nov 23 21:04:23 crc kubenswrapper[4726]: E1123 21:04:23.321215 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a\": container with ID starting with 3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a not found: ID does not exist" containerID="3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.321300 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a"} err="failed to get container status \"3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a\": rpc error: code = NotFound desc = could not find container \"3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a\": container with ID starting with 3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a not found: ID does not exist" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.321397 4726 scope.go:117] "RemoveContainer" containerID="0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1" Nov 23 21:04:23 crc kubenswrapper[4726]: E1123 21:04:23.321816 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1\": container with ID starting with 0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1 not found: ID does not exist" containerID="0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1" Nov 23 21:04:23 crc kubenswrapper[4726]: I1123 21:04:23.321934 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1"} err="failed to get container status \"0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1\": rpc error: code = NotFound desc = could not find container \"0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1\": container with ID starting with 0b22819cbc7444946ac4ccc1cf0996f741a1d3d403319db70d12cbc2c2688ae1 not found: ID does not exist" Nov 23 21:04:24 crc kubenswrapper[4726]: I1123 21:04:24.602491 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" path="/var/lib/kubelet/pods/3bccb5da-c358-49f7-bc96-478badba2d64/volumes" Nov 23 21:04:26 crc kubenswrapper[4726]: E1123 21:04:26.108339 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-conmon-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache]" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.054818 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mkn6p"] Nov 23 21:04:34 crc kubenswrapper[4726]: E1123 21:04:34.055695 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" containerName="extract-utilities" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.055708 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" containerName="extract-utilities" Nov 23 21:04:34 crc kubenswrapper[4726]: E1123 21:04:34.055720 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="extract-content" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.055725 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="extract-content" Nov 23 21:04:34 crc kubenswrapper[4726]: E1123 21:04:34.055739 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="extract-utilities" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.055745 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="extract-utilities" Nov 23 21:04:34 crc kubenswrapper[4726]: E1123 21:04:34.055757 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" containerName="registry-server" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.055764 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" containerName="registry-server" Nov 23 21:04:34 crc kubenswrapper[4726]: E1123 21:04:34.055771 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="registry-server" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.055776 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="registry-server" Nov 23 21:04:34 crc kubenswrapper[4726]: E1123 21:04:34.055799 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" containerName="extract-content" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.055805 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" containerName="extract-content" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.056014 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb4b648-4404-4a69-a6b5-9f8c7dd0973c" containerName="registry-server" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.056032 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bccb5da-c358-49f7-bc96-478badba2d64" containerName="registry-server" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.057262 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.068931 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mkn6p"] Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.204349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpgzn\" (UniqueName: \"kubernetes.io/projected/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-kube-api-access-zpgzn\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.204450 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-utilities\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.204526 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-catalog-content\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.306029 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpgzn\" (UniqueName: \"kubernetes.io/projected/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-kube-api-access-zpgzn\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.306093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-utilities\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.306124 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-catalog-content\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.306689 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-utilities\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.306745 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-catalog-content\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.331651 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpgzn\" (UniqueName: \"kubernetes.io/projected/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-kube-api-access-zpgzn\") pod \"community-operators-mkn6p\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.378320 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:34 crc kubenswrapper[4726]: I1123 21:04:34.977375 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mkn6p"] Nov 23 21:04:35 crc kubenswrapper[4726]: I1123 21:04:35.319224 4726 generic.go:334] "Generic (PLEG): container finished" podID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerID="df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020" exitCode=0 Nov 23 21:04:35 crc kubenswrapper[4726]: I1123 21:04:35.319289 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkn6p" event={"ID":"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c","Type":"ContainerDied","Data":"df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020"} Nov 23 21:04:35 crc kubenswrapper[4726]: I1123 21:04:35.319478 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkn6p" event={"ID":"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c","Type":"ContainerStarted","Data":"e615bc749cea1772e2d6767a05aad5161cc4599aa318690aac23ac643fabe30c"} Nov 23 21:04:36 crc kubenswrapper[4726]: I1123 21:04:36.331602 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkn6p" event={"ID":"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c","Type":"ContainerStarted","Data":"bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01"} Nov 23 21:04:36 crc kubenswrapper[4726]: E1123 21:04:36.364043 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-conmon-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache]" Nov 23 21:04:37 crc kubenswrapper[4726]: I1123 21:04:37.340858 4726 generic.go:334] "Generic (PLEG): container finished" podID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerID="bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01" exitCode=0 Nov 23 21:04:37 crc kubenswrapper[4726]: I1123 21:04:37.341209 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkn6p" event={"ID":"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c","Type":"ContainerDied","Data":"bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01"} Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.352483 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkn6p" event={"ID":"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c","Type":"ContainerStarted","Data":"893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0"} Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.378259 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mkn6p" podStartSLOduration=1.9411737470000001 podStartE2EDuration="4.378240397s" podCreationTimestamp="2025-11-23 21:04:34 +0000 UTC" firstStartedPulling="2025-11-23 21:04:35.320831194 +0000 UTC m=+3383.469872150" lastFinishedPulling="2025-11-23 21:04:37.757897814 +0000 UTC m=+3385.906938800" observedRunningTime="2025-11-23 21:04:38.369294807 +0000 UTC m=+3386.518335763" watchObservedRunningTime="2025-11-23 21:04:38.378240397 +0000 UTC m=+3386.527281353" Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.869697 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.871166 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.877488 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.877554 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-xbl4l" Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.877832 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.878006 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 23 21:04:38 crc kubenswrapper[4726]: I1123 21:04:38.892916 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011067 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011199 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011282 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011331 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pcxg\" (UniqueName: \"kubernetes.io/projected/efb73145-b382-4a15-89fe-6b74bc9d3594-kube-api-access-8pcxg\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011364 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011431 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011476 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011500 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.011626 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-config-data\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.043341 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.043402 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.043448 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.044213 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ce91d02edffb0448f1e3ab904f83011317aa44215088577adc4f62202e8da65"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.044269 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://8ce91d02edffb0448f1e3ab904f83011317aa44215088577adc4f62202e8da65" gracePeriod=600 Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.112992 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-config-data\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113121 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113152 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113206 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113240 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pcxg\" (UniqueName: \"kubernetes.io/projected/efb73145-b382-4a15-89fe-6b74bc9d3594-kube-api-access-8pcxg\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113273 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113315 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113347 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113374 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.113996 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.114270 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-config-data\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.114260 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.114606 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.114660 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.118970 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.120976 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.123914 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.131755 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pcxg\" (UniqueName: \"kubernetes.io/projected/efb73145-b382-4a15-89fe-6b74bc9d3594-kube-api-access-8pcxg\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.146666 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.187753 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.367371 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="8ce91d02edffb0448f1e3ab904f83011317aa44215088577adc4f62202e8da65" exitCode=0 Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.368143 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"8ce91d02edffb0448f1e3ab904f83011317aa44215088577adc4f62202e8da65"} Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.368193 4726 scope.go:117] "RemoveContainer" containerID="b606c3a13d50ced979c0adf5943ba809871613a9c1d5084b80dbab88f78568b4" Nov 23 21:04:39 crc kubenswrapper[4726]: I1123 21:04:39.671730 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 23 21:04:40 crc kubenswrapper[4726]: I1123 21:04:40.378845 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"efb73145-b382-4a15-89fe-6b74bc9d3594","Type":"ContainerStarted","Data":"871bee6dc4b1effc188424a1303bdab763b193bdfc728d8a7dafd29fce604fcc"} Nov 23 21:04:40 crc kubenswrapper[4726]: I1123 21:04:40.384671 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5"} Nov 23 21:04:44 crc kubenswrapper[4726]: I1123 21:04:44.379223 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:44 crc kubenswrapper[4726]: I1123 21:04:44.379600 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:44 crc kubenswrapper[4726]: I1123 21:04:44.437016 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:44 crc kubenswrapper[4726]: I1123 21:04:44.489806 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:44 crc kubenswrapper[4726]: I1123 21:04:44.681076 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mkn6p"] Nov 23 21:04:46 crc kubenswrapper[4726]: I1123 21:04:46.451039 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mkn6p" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerName="registry-server" containerID="cri-o://893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0" gracePeriod=2 Nov 23 21:04:46 crc kubenswrapper[4726]: E1123 21:04:46.739728 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-conmon-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02cf4ebc_aa8c_44fd_b933_05a6f52c2a8c.slice/crio-893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02cf4ebc_aa8c_44fd_b933_05a6f52c2a8c.slice/crio-conmon-893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0.scope\": RecentStats: unable to find data in memory cache]" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.232057 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.396258 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpgzn\" (UniqueName: \"kubernetes.io/projected/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-kube-api-access-zpgzn\") pod \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.396346 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-catalog-content\") pod \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.396457 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-utilities\") pod \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\" (UID: \"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c\") " Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.397378 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-utilities" (OuterVolumeSpecName: "utilities") pod "02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" (UID: "02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.402730 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-kube-api-access-zpgzn" (OuterVolumeSpecName: "kube-api-access-zpgzn") pod "02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" (UID: "02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c"). InnerVolumeSpecName "kube-api-access-zpgzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.448401 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" (UID: "02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.459526 4726 generic.go:334] "Generic (PLEG): container finished" podID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerID="893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0" exitCode=0 Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.459565 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkn6p" event={"ID":"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c","Type":"ContainerDied","Data":"893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0"} Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.459598 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mkn6p" event={"ID":"02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c","Type":"ContainerDied","Data":"e615bc749cea1772e2d6767a05aad5161cc4599aa318690aac23ac643fabe30c"} Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.459617 4726 scope.go:117] "RemoveContainer" containerID="893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.459619 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mkn6p" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.480174 4726 scope.go:117] "RemoveContainer" containerID="bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.499289 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpgzn\" (UniqueName: \"kubernetes.io/projected/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-kube-api-access-zpgzn\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.499320 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.499329 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.502541 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mkn6p"] Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.509330 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mkn6p"] Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.523146 4726 scope.go:117] "RemoveContainer" containerID="df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.554473 4726 scope.go:117] "RemoveContainer" containerID="893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0" Nov 23 21:04:47 crc kubenswrapper[4726]: E1123 21:04:47.554832 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0\": container with ID starting with 893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0 not found: ID does not exist" containerID="893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.554895 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0"} err="failed to get container status \"893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0\": rpc error: code = NotFound desc = could not find container \"893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0\": container with ID starting with 893f4b9ace176ef53ceef4ef260b77ab2cb89c7260973b6cf54620d93b61e1d0 not found: ID does not exist" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.554922 4726 scope.go:117] "RemoveContainer" containerID="bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01" Nov 23 21:04:47 crc kubenswrapper[4726]: E1123 21:04:47.555192 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01\": container with ID starting with bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01 not found: ID does not exist" containerID="bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.555220 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01"} err="failed to get container status \"bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01\": rpc error: code = NotFound desc = could not find container \"bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01\": container with ID starting with bd2d8dbe06e6a83cfb074a9cb5d343298e2fa14e1283259ef7c789c84d8d4f01 not found: ID does not exist" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.555240 4726 scope.go:117] "RemoveContainer" containerID="df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020" Nov 23 21:04:47 crc kubenswrapper[4726]: E1123 21:04:47.555395 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020\": container with ID starting with df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020 not found: ID does not exist" containerID="df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020" Nov 23 21:04:47 crc kubenswrapper[4726]: I1123 21:04:47.555421 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020"} err="failed to get container status \"df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020\": rpc error: code = NotFound desc = could not find container \"df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020\": container with ID starting with df560f81b6113b533b0f65d0a35bc092ae62aa21089183a556171d7dd3912020 not found: ID does not exist" Nov 23 21:04:48 crc kubenswrapper[4726]: I1123 21:04:48.601447 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" path="/var/lib/kubelet/pods/02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c/volumes" Nov 23 21:04:56 crc kubenswrapper[4726]: E1123 21:04:56.984672 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-conmon-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache]" Nov 23 21:05:07 crc kubenswrapper[4726]: E1123 21:05:07.519852 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bccb5da_c358_49f7_bc96_478badba2d64.slice/crio-conmon-3adb1b146c2df5ee226f7057aba7b816fa829e97a8fde265ca1e61d16411239a.scope\": RecentStats: unable to find data in memory cache]" Nov 23 21:05:12 crc kubenswrapper[4726]: E1123 21:05:12.619998 4726 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/9c2115468ee4983534aa3fad90449c45b3550bc86b32bddd2d230320116a7608/diff" to get inode usage: stat /var/lib/containers/storage/overlay/9c2115468ee4983534aa3fad90449c45b3550bc86b32bddd2d230320116a7608/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-marketplace_redhat-marketplace-t7hfp_3bccb5da-c358-49f7-bc96-478badba2d64/extract-content/0.log" to get inode usage: stat /var/log/pods/openshift-marketplace_redhat-marketplace-t7hfp_3bccb5da-c358-49f7-bc96-478badba2d64/extract-content/0.log: no such file or directory Nov 23 21:05:16 crc kubenswrapper[4726]: E1123 21:05:16.405362 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 23 21:05:16 crc kubenswrapper[4726]: E1123 21:05:16.412381 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8pcxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(efb73145-b382-4a15-89fe-6b74bc9d3594): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 23 21:05:16 crc kubenswrapper[4726]: E1123 21:05:16.413555 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="efb73145-b382-4a15-89fe-6b74bc9d3594" Nov 23 21:05:16 crc kubenswrapper[4726]: E1123 21:05:16.727256 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="efb73145-b382-4a15-89fe-6b74bc9d3594" Nov 23 21:05:31 crc kubenswrapper[4726]: I1123 21:05:31.002152 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 23 21:05:32 crc kubenswrapper[4726]: I1123 21:05:32.913283 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"efb73145-b382-4a15-89fe-6b74bc9d3594","Type":"ContainerStarted","Data":"4a5f659c08fe10cf885efd9a8c84644882d383875e496e0932c27be349b3a8fc"} Nov 23 21:05:32 crc kubenswrapper[4726]: I1123 21:05:32.945020 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.635506664 podStartE2EDuration="55.944991815s" podCreationTimestamp="2025-11-23 21:04:37 +0000 UTC" firstStartedPulling="2025-11-23 21:04:39.685550193 +0000 UTC m=+3387.834591149" lastFinishedPulling="2025-11-23 21:05:30.995035344 +0000 UTC m=+3439.144076300" observedRunningTime="2025-11-23 21:05:32.935204562 +0000 UTC m=+3441.084245528" watchObservedRunningTime="2025-11-23 21:05:32.944991815 +0000 UTC m=+3441.094032811" Nov 23 21:06:39 crc kubenswrapper[4726]: I1123 21:06:39.043112 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:06:39 crc kubenswrapper[4726]: I1123 21:06:39.043671 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:07:09 crc kubenswrapper[4726]: I1123 21:07:09.043164 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:07:09 crc kubenswrapper[4726]: I1123 21:07:09.043667 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.042797 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.043377 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.043422 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.044164 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.044219 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" gracePeriod=600 Nov 23 21:07:39 crc kubenswrapper[4726]: E1123 21:07:39.174483 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.193429 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" exitCode=0 Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.193477 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5"} Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.193511 4726 scope.go:117] "RemoveContainer" containerID="8ce91d02edffb0448f1e3ab904f83011317aa44215088577adc4f62202e8da65" Nov 23 21:07:39 crc kubenswrapper[4726]: I1123 21:07:39.195377 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:07:39 crc kubenswrapper[4726]: E1123 21:07:39.196713 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:07:49 crc kubenswrapper[4726]: I1123 21:07:49.589514 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:07:49 crc kubenswrapper[4726]: E1123 21:07:49.590299 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:08:04 crc kubenswrapper[4726]: I1123 21:08:04.589582 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:08:04 crc kubenswrapper[4726]: E1123 21:08:04.590528 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:08:17 crc kubenswrapper[4726]: I1123 21:08:17.589397 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:08:17 crc kubenswrapper[4726]: E1123 21:08:17.590477 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:08:32 crc kubenswrapper[4726]: I1123 21:08:32.597723 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:08:32 crc kubenswrapper[4726]: E1123 21:08:32.598915 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:08:43 crc kubenswrapper[4726]: I1123 21:08:43.589424 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:08:43 crc kubenswrapper[4726]: E1123 21:08:43.590059 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:08:56 crc kubenswrapper[4726]: I1123 21:08:56.592387 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:08:56 crc kubenswrapper[4726]: E1123 21:08:56.593446 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:09:07 crc kubenswrapper[4726]: I1123 21:09:07.591647 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:09:07 crc kubenswrapper[4726]: E1123 21:09:07.592328 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:09:19 crc kubenswrapper[4726]: I1123 21:09:19.588699 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:09:19 crc kubenswrapper[4726]: E1123 21:09:19.589425 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:09:20 crc kubenswrapper[4726]: I1123 21:09:20.824328 4726 scope.go:117] "RemoveContainer" containerID="cb1aa9060bbe1832d9790d85beeb36b28fc3a098686ed8fd21fc16b4f3b05de0" Nov 23 21:09:20 crc kubenswrapper[4726]: I1123 21:09:20.869006 4726 scope.go:117] "RemoveContainer" containerID="8c18edecc8a6c983d4a337b1f1ead199ea7b235e4992b6f6eccc6c91ddeef922" Nov 23 21:09:30 crc kubenswrapper[4726]: I1123 21:09:30.589803 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:09:30 crc kubenswrapper[4726]: E1123 21:09:30.590741 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:09:42 crc kubenswrapper[4726]: I1123 21:09:42.599075 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:09:42 crc kubenswrapper[4726]: E1123 21:09:42.600093 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:09:55 crc kubenswrapper[4726]: I1123 21:09:55.590691 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:09:55 crc kubenswrapper[4726]: E1123 21:09:55.591562 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:10:06 crc kubenswrapper[4726]: I1123 21:10:06.589633 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:10:06 crc kubenswrapper[4726]: E1123 21:10:06.590572 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:10:19 crc kubenswrapper[4726]: I1123 21:10:19.590309 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:10:19 crc kubenswrapper[4726]: E1123 21:10:19.591100 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:10:34 crc kubenswrapper[4726]: I1123 21:10:34.589650 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:10:34 crc kubenswrapper[4726]: E1123 21:10:34.590423 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:10:49 crc kubenswrapper[4726]: I1123 21:10:49.589776 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:10:49 crc kubenswrapper[4726]: E1123 21:10:49.590611 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:11:03 crc kubenswrapper[4726]: I1123 21:11:03.589744 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:11:03 crc kubenswrapper[4726]: E1123 21:11:03.590703 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:11:08 crc kubenswrapper[4726]: I1123 21:11:08.789423 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="e88e0fb1-897d-454c-9a23-8187592296e5" containerName="galera" probeResult="failure" output="command timed out" Nov 23 21:11:18 crc kubenswrapper[4726]: I1123 21:11:18.589851 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:11:18 crc kubenswrapper[4726]: E1123 21:11:18.590824 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:11:32 crc kubenswrapper[4726]: I1123 21:11:32.605725 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:11:32 crc kubenswrapper[4726]: E1123 21:11:32.606651 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:11:43 crc kubenswrapper[4726]: I1123 21:11:43.589438 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:11:43 crc kubenswrapper[4726]: E1123 21:11:43.591266 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:11:58 crc kubenswrapper[4726]: I1123 21:11:58.589793 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:11:58 crc kubenswrapper[4726]: E1123 21:11:58.590775 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:12:13 crc kubenswrapper[4726]: I1123 21:12:13.589418 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:12:13 crc kubenswrapper[4726]: E1123 21:12:13.590220 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:12:20 crc kubenswrapper[4726]: I1123 21:12:20.071501 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-2cfvg"] Nov 23 21:12:20 crc kubenswrapper[4726]: I1123 21:12:20.089573 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-2cfvg"] Nov 23 21:12:20 crc kubenswrapper[4726]: I1123 21:12:20.608318 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf6e3f28-0a31-4cb0-8213-988ce169be1a" path="/var/lib/kubelet/pods/cf6e3f28-0a31-4cb0-8213-988ce169be1a/volumes" Nov 23 21:12:20 crc kubenswrapper[4726]: I1123 21:12:20.972238 4726 scope.go:117] "RemoveContainer" containerID="ac651bac234623357112a86a3868f9d9d20f2142f4cec3b7348ab1bdf61b5435" Nov 23 21:12:21 crc kubenswrapper[4726]: I1123 21:12:21.045589 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-c3e6-account-create-fk69w"] Nov 23 21:12:21 crc kubenswrapper[4726]: I1123 21:12:21.054859 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-c3e6-account-create-fk69w"] Nov 23 21:12:22 crc kubenswrapper[4726]: I1123 21:12:22.609854 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8e3f5ad-3617-40ce-9053-a32c9d2cde1c" path="/var/lib/kubelet/pods/a8e3f5ad-3617-40ce-9053-a32c9d2cde1c/volumes" Nov 23 21:12:24 crc kubenswrapper[4726]: I1123 21:12:24.590747 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:12:24 crc kubenswrapper[4726]: E1123 21:12:24.591414 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.059109 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t7brq"] Nov 23 21:12:33 crc kubenswrapper[4726]: E1123 21:12:33.063235 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerName="registry-server" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.063359 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerName="registry-server" Nov 23 21:12:33 crc kubenswrapper[4726]: E1123 21:12:33.063447 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerName="extract-content" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.063523 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerName="extract-content" Nov 23 21:12:33 crc kubenswrapper[4726]: E1123 21:12:33.063609 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerName="extract-utilities" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.063686 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerName="extract-utilities" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.063988 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="02cf4ebc-aa8c-44fd-b933-05a6f52c2a8c" containerName="registry-server" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.065997 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.074128 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t7brq"] Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.142851 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-catalog-content\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.143004 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-utilities\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.143080 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tch8z\" (UniqueName: \"kubernetes.io/projected/82f71ad7-c259-44fe-b1e4-1d01b30c241b-kube-api-access-tch8z\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.245791 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-utilities\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.246094 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tch8z\" (UniqueName: \"kubernetes.io/projected/82f71ad7-c259-44fe-b1e4-1d01b30c241b-kube-api-access-tch8z\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.246295 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-catalog-content\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.247229 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-utilities\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.247937 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-catalog-content\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.276971 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tch8z\" (UniqueName: \"kubernetes.io/projected/82f71ad7-c259-44fe-b1e4-1d01b30c241b-kube-api-access-tch8z\") pod \"certified-operators-t7brq\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:33 crc kubenswrapper[4726]: I1123 21:12:33.420104 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:34 crc kubenswrapper[4726]: I1123 21:12:34.333089 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t7brq"] Nov 23 21:12:34 crc kubenswrapper[4726]: W1123 21:12:34.355020 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82f71ad7_c259_44fe_b1e4_1d01b30c241b.slice/crio-3515cfc459a3a952e45c24983938684bf248515fda48128b510e2b2bbd1267bf WatchSource:0}: Error finding container 3515cfc459a3a952e45c24983938684bf248515fda48128b510e2b2bbd1267bf: Status 404 returned error can't find the container with id 3515cfc459a3a952e45c24983938684bf248515fda48128b510e2b2bbd1267bf Nov 23 21:12:35 crc kubenswrapper[4726]: I1123 21:12:35.188907 4726 generic.go:334] "Generic (PLEG): container finished" podID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerID="f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a" exitCode=0 Nov 23 21:12:35 crc kubenswrapper[4726]: I1123 21:12:35.189007 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7brq" event={"ID":"82f71ad7-c259-44fe-b1e4-1d01b30c241b","Type":"ContainerDied","Data":"f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a"} Nov 23 21:12:35 crc kubenswrapper[4726]: I1123 21:12:35.189621 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7brq" event={"ID":"82f71ad7-c259-44fe-b1e4-1d01b30c241b","Type":"ContainerStarted","Data":"3515cfc459a3a952e45c24983938684bf248515fda48128b510e2b2bbd1267bf"} Nov 23 21:12:35 crc kubenswrapper[4726]: I1123 21:12:35.198862 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 21:12:36 crc kubenswrapper[4726]: I1123 21:12:36.200482 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7brq" event={"ID":"82f71ad7-c259-44fe-b1e4-1d01b30c241b","Type":"ContainerStarted","Data":"2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9"} Nov 23 21:12:37 crc kubenswrapper[4726]: I1123 21:12:37.590832 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:12:37 crc kubenswrapper[4726]: E1123 21:12:37.591426 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:12:38 crc kubenswrapper[4726]: I1123 21:12:38.225685 4726 generic.go:334] "Generic (PLEG): container finished" podID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerID="2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9" exitCode=0 Nov 23 21:12:38 crc kubenswrapper[4726]: I1123 21:12:38.225757 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7brq" event={"ID":"82f71ad7-c259-44fe-b1e4-1d01b30c241b","Type":"ContainerDied","Data":"2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9"} Nov 23 21:12:39 crc kubenswrapper[4726]: I1123 21:12:39.237727 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7brq" event={"ID":"82f71ad7-c259-44fe-b1e4-1d01b30c241b","Type":"ContainerStarted","Data":"5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc"} Nov 23 21:12:39 crc kubenswrapper[4726]: I1123 21:12:39.265368 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t7brq" podStartSLOduration=2.804073296 podStartE2EDuration="6.265347753s" podCreationTimestamp="2025-11-23 21:12:33 +0000 UTC" firstStartedPulling="2025-11-23 21:12:35.190797581 +0000 UTC m=+3863.339838547" lastFinishedPulling="2025-11-23 21:12:38.652072048 +0000 UTC m=+3866.801113004" observedRunningTime="2025-11-23 21:12:39.254979233 +0000 UTC m=+3867.404020199" watchObservedRunningTime="2025-11-23 21:12:39.265347753 +0000 UTC m=+3867.414388719" Nov 23 21:12:43 crc kubenswrapper[4726]: I1123 21:12:43.420943 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:43 crc kubenswrapper[4726]: I1123 21:12:43.421776 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:44 crc kubenswrapper[4726]: I1123 21:12:44.484617 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-t7brq" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="registry-server" probeResult="failure" output=< Nov 23 21:12:44 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 21:12:44 crc kubenswrapper[4726]: > Nov 23 21:12:48 crc kubenswrapper[4726]: I1123 21:12:48.589145 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:12:49 crc kubenswrapper[4726]: I1123 21:12:49.326303 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"7af01d69b4d8696a9058f0f4eaa0367543076fc2495dc5f83abbcf020b3a86d7"} Nov 23 21:12:51 crc kubenswrapper[4726]: I1123 21:12:51.067265 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-6bxcm"] Nov 23 21:12:51 crc kubenswrapper[4726]: I1123 21:12:51.078332 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-6bxcm"] Nov 23 21:12:52 crc kubenswrapper[4726]: I1123 21:12:52.607262 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b" path="/var/lib/kubelet/pods/4d583328-dc5f-49fd-b72d-c7dfbcbf1a3b/volumes" Nov 23 21:12:53 crc kubenswrapper[4726]: I1123 21:12:53.492004 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:53 crc kubenswrapper[4726]: I1123 21:12:53.570731 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:53 crc kubenswrapper[4726]: I1123 21:12:53.735013 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t7brq"] Nov 23 21:12:55 crc kubenswrapper[4726]: I1123 21:12:55.380395 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t7brq" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="registry-server" containerID="cri-o://5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc" gracePeriod=2 Nov 23 21:12:55 crc kubenswrapper[4726]: I1123 21:12:55.970784 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.117481 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-catalog-content\") pod \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.118158 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-utilities\") pod \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.119471 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-utilities" (OuterVolumeSpecName: "utilities") pod "82f71ad7-c259-44fe-b1e4-1d01b30c241b" (UID: "82f71ad7-c259-44fe-b1e4-1d01b30c241b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.119732 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tch8z\" (UniqueName: \"kubernetes.io/projected/82f71ad7-c259-44fe-b1e4-1d01b30c241b-kube-api-access-tch8z\") pod \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\" (UID: \"82f71ad7-c259-44fe-b1e4-1d01b30c241b\") " Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.124441 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.127533 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82f71ad7-c259-44fe-b1e4-1d01b30c241b-kube-api-access-tch8z" (OuterVolumeSpecName: "kube-api-access-tch8z") pod "82f71ad7-c259-44fe-b1e4-1d01b30c241b" (UID: "82f71ad7-c259-44fe-b1e4-1d01b30c241b"). InnerVolumeSpecName "kube-api-access-tch8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.160493 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82f71ad7-c259-44fe-b1e4-1d01b30c241b" (UID: "82f71ad7-c259-44fe-b1e4-1d01b30c241b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.226507 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tch8z\" (UniqueName: \"kubernetes.io/projected/82f71ad7-c259-44fe-b1e4-1d01b30c241b-kube-api-access-tch8z\") on node \"crc\" DevicePath \"\"" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.226542 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f71ad7-c259-44fe-b1e4-1d01b30c241b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.389369 4726 generic.go:334] "Generic (PLEG): container finished" podID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerID="5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc" exitCode=0 Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.389404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7brq" event={"ID":"82f71ad7-c259-44fe-b1e4-1d01b30c241b","Type":"ContainerDied","Data":"5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc"} Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.389434 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7brq" event={"ID":"82f71ad7-c259-44fe-b1e4-1d01b30c241b","Type":"ContainerDied","Data":"3515cfc459a3a952e45c24983938684bf248515fda48128b510e2b2bbd1267bf"} Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.389445 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7brq" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.389450 4726 scope.go:117] "RemoveContainer" containerID="5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.414796 4726 scope.go:117] "RemoveContainer" containerID="2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.424071 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t7brq"] Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.434317 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t7brq"] Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.445992 4726 scope.go:117] "RemoveContainer" containerID="f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.479473 4726 scope.go:117] "RemoveContainer" containerID="5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc" Nov 23 21:12:56 crc kubenswrapper[4726]: E1123 21:12:56.479959 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc\": container with ID starting with 5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc not found: ID does not exist" containerID="5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.479996 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc"} err="failed to get container status \"5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc\": rpc error: code = NotFound desc = could not find container \"5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc\": container with ID starting with 5840b5c4e2ee4468382fde06acb550f7609a97251cbeed42e6a47259c76ae9bc not found: ID does not exist" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.480026 4726 scope.go:117] "RemoveContainer" containerID="2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9" Nov 23 21:12:56 crc kubenswrapper[4726]: E1123 21:12:56.480335 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9\": container with ID starting with 2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9 not found: ID does not exist" containerID="2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.480381 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9"} err="failed to get container status \"2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9\": rpc error: code = NotFound desc = could not find container \"2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9\": container with ID starting with 2c27d5782d9f11627a35e7371e8218ddf67e8beef13dc6795a84db0629a102d9 not found: ID does not exist" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.480412 4726 scope.go:117] "RemoveContainer" containerID="f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a" Nov 23 21:12:56 crc kubenswrapper[4726]: E1123 21:12:56.480713 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a\": container with ID starting with f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a not found: ID does not exist" containerID="f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.480735 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a"} err="failed to get container status \"f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a\": rpc error: code = NotFound desc = could not find container \"f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a\": container with ID starting with f8e4b623dbc7108350597a2eea71b370606325b6caa4e770fff902e5a7d3462a not found: ID does not exist" Nov 23 21:12:56 crc kubenswrapper[4726]: I1123 21:12:56.599403 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" path="/var/lib/kubelet/pods/82f71ad7-c259-44fe-b1e4-1d01b30c241b/volumes" Nov 23 21:13:21 crc kubenswrapper[4726]: I1123 21:13:21.081087 4726 scope.go:117] "RemoveContainer" containerID="5627ac8f7118a0b69cf2b9646826d7b84145f83703783f18735d6f43c79b2899" Nov 23 21:13:21 crc kubenswrapper[4726]: I1123 21:13:21.107795 4726 scope.go:117] "RemoveContainer" containerID="ed3639f6be29fd4e8cb8338756f5c41b31bb7a8f749a9a72c2160cd2151346c1" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.658257 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fpcfn"] Nov 23 21:14:37 crc kubenswrapper[4726]: E1123 21:14:37.659480 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="extract-content" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.659506 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="extract-content" Nov 23 21:14:37 crc kubenswrapper[4726]: E1123 21:14:37.659530 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="registry-server" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.659541 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="registry-server" Nov 23 21:14:37 crc kubenswrapper[4726]: E1123 21:14:37.659578 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="extract-utilities" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.659586 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="extract-utilities" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.659841 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="82f71ad7-c259-44fe-b1e4-1d01b30c241b" containerName="registry-server" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.661659 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.672683 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fpcfn"] Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.763580 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-utilities\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.763661 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhrj5\" (UniqueName: \"kubernetes.io/projected/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-kube-api-access-dhrj5\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.763892 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-catalog-content\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.866417 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-utilities\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.866505 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhrj5\" (UniqueName: \"kubernetes.io/projected/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-kube-api-access-dhrj5\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.866546 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-catalog-content\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.867154 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-catalog-content\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.867289 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-utilities\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.887322 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhrj5\" (UniqueName: \"kubernetes.io/projected/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-kube-api-access-dhrj5\") pod \"redhat-operators-fpcfn\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:37 crc kubenswrapper[4726]: I1123 21:14:37.998626 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:38 crc kubenswrapper[4726]: I1123 21:14:38.511317 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fpcfn"] Nov 23 21:14:39 crc kubenswrapper[4726]: I1123 21:14:39.398619 4726 generic.go:334] "Generic (PLEG): container finished" podID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerID="6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6" exitCode=0 Nov 23 21:14:39 crc kubenswrapper[4726]: I1123 21:14:39.398691 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpcfn" event={"ID":"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc","Type":"ContainerDied","Data":"6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6"} Nov 23 21:14:39 crc kubenswrapper[4726]: I1123 21:14:39.399025 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpcfn" event={"ID":"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc","Type":"ContainerStarted","Data":"b6e94cf90341006336a0fb51f12e1568b95f99e55569879342bc62925e56e2db"} Nov 23 21:14:40 crc kubenswrapper[4726]: I1123 21:14:40.411602 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpcfn" event={"ID":"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc","Type":"ContainerStarted","Data":"6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19"} Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.462025 4726 generic.go:334] "Generic (PLEG): container finished" podID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerID="6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19" exitCode=0 Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.462109 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpcfn" event={"ID":"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc","Type":"ContainerDied","Data":"6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19"} Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.718005 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w568x"] Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.722124 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.743209 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w568x"] Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.873100 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mv4b\" (UniqueName: \"kubernetes.io/projected/d88865da-26c0-4533-81f1-f1b311a00739-kube-api-access-4mv4b\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.873674 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-catalog-content\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.873718 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-utilities\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.976663 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-catalog-content\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.976862 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-utilities\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.977256 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mv4b\" (UniqueName: \"kubernetes.io/projected/d88865da-26c0-4533-81f1-f1b311a00739-kube-api-access-4mv4b\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.980692 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-utilities\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:44 crc kubenswrapper[4726]: I1123 21:14:44.981153 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-catalog-content\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:45 crc kubenswrapper[4726]: I1123 21:14:45.006534 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mv4b\" (UniqueName: \"kubernetes.io/projected/d88865da-26c0-4533-81f1-f1b311a00739-kube-api-access-4mv4b\") pod \"community-operators-w568x\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:45 crc kubenswrapper[4726]: I1123 21:14:45.045891 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:45 crc kubenswrapper[4726]: I1123 21:14:45.952453 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w568x"] Nov 23 21:14:46 crc kubenswrapper[4726]: I1123 21:14:46.485609 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpcfn" event={"ID":"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc","Type":"ContainerStarted","Data":"da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684"} Nov 23 21:14:46 crc kubenswrapper[4726]: I1123 21:14:46.487607 4726 generic.go:334] "Generic (PLEG): container finished" podID="d88865da-26c0-4533-81f1-f1b311a00739" containerID="58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161" exitCode=0 Nov 23 21:14:46 crc kubenswrapper[4726]: I1123 21:14:46.487670 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w568x" event={"ID":"d88865da-26c0-4533-81f1-f1b311a00739","Type":"ContainerDied","Data":"58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161"} Nov 23 21:14:46 crc kubenswrapper[4726]: I1123 21:14:46.487709 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w568x" event={"ID":"d88865da-26c0-4533-81f1-f1b311a00739","Type":"ContainerStarted","Data":"00da91a3a65646177f4ba5c43989cd7d0e49e41c4e308b99491c4502803c49b3"} Nov 23 21:14:46 crc kubenswrapper[4726]: I1123 21:14:46.510702 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fpcfn" podStartSLOduration=3.793130966 podStartE2EDuration="9.510680741s" podCreationTimestamp="2025-11-23 21:14:37 +0000 UTC" firstStartedPulling="2025-11-23 21:14:39.401807973 +0000 UTC m=+3987.550848919" lastFinishedPulling="2025-11-23 21:14:45.119357748 +0000 UTC m=+3993.268398694" observedRunningTime="2025-11-23 21:14:46.504173298 +0000 UTC m=+3994.653214254" watchObservedRunningTime="2025-11-23 21:14:46.510680741 +0000 UTC m=+3994.659721697" Nov 23 21:14:48 crc kubenswrapper[4726]: I1123 21:14:47.999313 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:48 crc kubenswrapper[4726]: I1123 21:14:48.000022 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:14:48 crc kubenswrapper[4726]: I1123 21:14:48.534154 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w568x" event={"ID":"d88865da-26c0-4533-81f1-f1b311a00739","Type":"ContainerStarted","Data":"19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374"} Nov 23 21:14:49 crc kubenswrapper[4726]: I1123 21:14:49.063918 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fpcfn" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="registry-server" probeResult="failure" output=< Nov 23 21:14:49 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 21:14:49 crc kubenswrapper[4726]: > Nov 23 21:14:49 crc kubenswrapper[4726]: I1123 21:14:49.546386 4726 generic.go:334] "Generic (PLEG): container finished" podID="d88865da-26c0-4533-81f1-f1b311a00739" containerID="19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374" exitCode=0 Nov 23 21:14:49 crc kubenswrapper[4726]: I1123 21:14:49.546916 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w568x" event={"ID":"d88865da-26c0-4533-81f1-f1b311a00739","Type":"ContainerDied","Data":"19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374"} Nov 23 21:14:50 crc kubenswrapper[4726]: I1123 21:14:50.557757 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w568x" event={"ID":"d88865da-26c0-4533-81f1-f1b311a00739","Type":"ContainerStarted","Data":"ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720"} Nov 23 21:14:50 crc kubenswrapper[4726]: I1123 21:14:50.580649 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w568x" podStartSLOduration=3.128050454 podStartE2EDuration="6.580629542s" podCreationTimestamp="2025-11-23 21:14:44 +0000 UTC" firstStartedPulling="2025-11-23 21:14:46.489927858 +0000 UTC m=+3994.638968814" lastFinishedPulling="2025-11-23 21:14:49.942506936 +0000 UTC m=+3998.091547902" observedRunningTime="2025-11-23 21:14:50.577351239 +0000 UTC m=+3998.726392195" watchObservedRunningTime="2025-11-23 21:14:50.580629542 +0000 UTC m=+3998.729670498" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.046297 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.046853 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.093666 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.504380 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fpsnb"] Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.506145 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.517995 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpsnb"] Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.615218 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-utilities\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.615291 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr8s6\" (UniqueName: \"kubernetes.io/projected/3af6c1aa-92a2-468a-837f-af8046c6db4f-kube-api-access-cr8s6\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.615323 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-catalog-content\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.657210 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.716624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-utilities\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.716678 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr8s6\" (UniqueName: \"kubernetes.io/projected/3af6c1aa-92a2-468a-837f-af8046c6db4f-kube-api-access-cr8s6\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.716708 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-catalog-content\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.717157 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-utilities\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.717191 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-catalog-content\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.734127 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr8s6\" (UniqueName: \"kubernetes.io/projected/3af6c1aa-92a2-468a-837f-af8046c6db4f-kube-api-access-cr8s6\") pod \"redhat-marketplace-fpsnb\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:55 crc kubenswrapper[4726]: I1123 21:14:55.826154 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:14:56 crc kubenswrapper[4726]: I1123 21:14:56.322408 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpsnb"] Nov 23 21:14:56 crc kubenswrapper[4726]: W1123 21:14:56.327878 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3af6c1aa_92a2_468a_837f_af8046c6db4f.slice/crio-e8bdcc21650d64e2980e0a85114d0417cd54325dad57a4c6126324ab9f69fedf WatchSource:0}: Error finding container e8bdcc21650d64e2980e0a85114d0417cd54325dad57a4c6126324ab9f69fedf: Status 404 returned error can't find the container with id e8bdcc21650d64e2980e0a85114d0417cd54325dad57a4c6126324ab9f69fedf Nov 23 21:14:56 crc kubenswrapper[4726]: I1123 21:14:56.615374 4726 generic.go:334] "Generic (PLEG): container finished" podID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerID="51a232cff9cd0e2817c67447b8205fbc735bc31d54ab965d13526ddf7519db2e" exitCode=0 Nov 23 21:14:56 crc kubenswrapper[4726]: I1123 21:14:56.615453 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpsnb" event={"ID":"3af6c1aa-92a2-468a-837f-af8046c6db4f","Type":"ContainerDied","Data":"51a232cff9cd0e2817c67447b8205fbc735bc31d54ab965d13526ddf7519db2e"} Nov 23 21:14:56 crc kubenswrapper[4726]: I1123 21:14:56.615719 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpsnb" event={"ID":"3af6c1aa-92a2-468a-837f-af8046c6db4f","Type":"ContainerStarted","Data":"e8bdcc21650d64e2980e0a85114d0417cd54325dad57a4c6126324ab9f69fedf"} Nov 23 21:14:57 crc kubenswrapper[4726]: I1123 21:14:57.628792 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpsnb" event={"ID":"3af6c1aa-92a2-468a-837f-af8046c6db4f","Type":"ContainerStarted","Data":"9c3e3fd52204d39ef481b804fc5cac3a9d03b6a328dec9f14ed20797b5ade051"} Nov 23 21:14:57 crc kubenswrapper[4726]: I1123 21:14:57.933165 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w568x"] Nov 23 21:14:57 crc kubenswrapper[4726]: I1123 21:14:57.933418 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w568x" podUID="d88865da-26c0-4533-81f1-f1b311a00739" containerName="registry-server" containerID="cri-o://ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720" gracePeriod=2 Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.426955 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.480764 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-utilities\") pod \"d88865da-26c0-4533-81f1-f1b311a00739\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.480870 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-catalog-content\") pod \"d88865da-26c0-4533-81f1-f1b311a00739\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.480966 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mv4b\" (UniqueName: \"kubernetes.io/projected/d88865da-26c0-4533-81f1-f1b311a00739-kube-api-access-4mv4b\") pod \"d88865da-26c0-4533-81f1-f1b311a00739\" (UID: \"d88865da-26c0-4533-81f1-f1b311a00739\") " Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.481991 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-utilities" (OuterVolumeSpecName: "utilities") pod "d88865da-26c0-4533-81f1-f1b311a00739" (UID: "d88865da-26c0-4533-81f1-f1b311a00739"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.488098 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d88865da-26c0-4533-81f1-f1b311a00739-kube-api-access-4mv4b" (OuterVolumeSpecName: "kube-api-access-4mv4b") pod "d88865da-26c0-4533-81f1-f1b311a00739" (UID: "d88865da-26c0-4533-81f1-f1b311a00739"). InnerVolumeSpecName "kube-api-access-4mv4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.535914 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d88865da-26c0-4533-81f1-f1b311a00739" (UID: "d88865da-26c0-4533-81f1-f1b311a00739"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.583050 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.583075 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d88865da-26c0-4533-81f1-f1b311a00739-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.583086 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mv4b\" (UniqueName: \"kubernetes.io/projected/d88865da-26c0-4533-81f1-f1b311a00739-kube-api-access-4mv4b\") on node \"crc\" DevicePath \"\"" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.640830 4726 generic.go:334] "Generic (PLEG): container finished" podID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerID="9c3e3fd52204d39ef481b804fc5cac3a9d03b6a328dec9f14ed20797b5ade051" exitCode=0 Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.640940 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpsnb" event={"ID":"3af6c1aa-92a2-468a-837f-af8046c6db4f","Type":"ContainerDied","Data":"9c3e3fd52204d39ef481b804fc5cac3a9d03b6a328dec9f14ed20797b5ade051"} Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.649693 4726 generic.go:334] "Generic (PLEG): container finished" podID="d88865da-26c0-4533-81f1-f1b311a00739" containerID="ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720" exitCode=0 Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.649733 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w568x" event={"ID":"d88865da-26c0-4533-81f1-f1b311a00739","Type":"ContainerDied","Data":"ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720"} Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.649735 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w568x" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.649759 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w568x" event={"ID":"d88865da-26c0-4533-81f1-f1b311a00739","Type":"ContainerDied","Data":"00da91a3a65646177f4ba5c43989cd7d0e49e41c4e308b99491c4502803c49b3"} Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.649777 4726 scope.go:117] "RemoveContainer" containerID="ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.679944 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w568x"] Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.687992 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w568x"] Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.690269 4726 scope.go:117] "RemoveContainer" containerID="19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.718633 4726 scope.go:117] "RemoveContainer" containerID="58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.756032 4726 scope.go:117] "RemoveContainer" containerID="ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720" Nov 23 21:14:58 crc kubenswrapper[4726]: E1123 21:14:58.756551 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720\": container with ID starting with ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720 not found: ID does not exist" containerID="ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.756692 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720"} err="failed to get container status \"ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720\": rpc error: code = NotFound desc = could not find container \"ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720\": container with ID starting with ce1fbafb80a38a155e129356c0ce1b926a3fc3f563d7b66f53eafaf1b29e6720 not found: ID does not exist" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.756821 4726 scope.go:117] "RemoveContainer" containerID="19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374" Nov 23 21:14:58 crc kubenswrapper[4726]: E1123 21:14:58.757292 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374\": container with ID starting with 19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374 not found: ID does not exist" containerID="19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.757324 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374"} err="failed to get container status \"19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374\": rpc error: code = NotFound desc = could not find container \"19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374\": container with ID starting with 19bfbb97b53c52278c8d64ab484c9895e6a8fd7cfb147cacbcbc32ab722c3374 not found: ID does not exist" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.757343 4726 scope.go:117] "RemoveContainer" containerID="58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161" Nov 23 21:14:58 crc kubenswrapper[4726]: E1123 21:14:58.757568 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161\": container with ID starting with 58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161 not found: ID does not exist" containerID="58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161" Nov 23 21:14:58 crc kubenswrapper[4726]: I1123 21:14:58.757592 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161"} err="failed to get container status \"58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161\": rpc error: code = NotFound desc = could not find container \"58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161\": container with ID starting with 58c78ecdc240411130a794afc682ac66c46b85261913db8231ed2e9273928161 not found: ID does not exist" Nov 23 21:14:59 crc kubenswrapper[4726]: I1123 21:14:59.044742 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fpcfn" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="registry-server" probeResult="failure" output=< Nov 23 21:14:59 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 21:14:59 crc kubenswrapper[4726]: > Nov 23 21:14:59 crc kubenswrapper[4726]: I1123 21:14:59.661004 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpsnb" event={"ID":"3af6c1aa-92a2-468a-837f-af8046c6db4f","Type":"ContainerStarted","Data":"68f10f5fc1b79051213dc60029b4fcaf18d2a161474dee448f0f6a8ddf828d74"} Nov 23 21:14:59 crc kubenswrapper[4726]: I1123 21:14:59.688074 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fpsnb" podStartSLOduration=2.242054149 podStartE2EDuration="4.688049521s" podCreationTimestamp="2025-11-23 21:14:55 +0000 UTC" firstStartedPulling="2025-11-23 21:14:56.617901396 +0000 UTC m=+4004.766942362" lastFinishedPulling="2025-11-23 21:14:59.063896768 +0000 UTC m=+4007.212937734" observedRunningTime="2025-11-23 21:14:59.684064809 +0000 UTC m=+4007.833105765" watchObservedRunningTime="2025-11-23 21:14:59.688049521 +0000 UTC m=+4007.837090497" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.172678 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2"] Nov 23 21:15:00 crc kubenswrapper[4726]: E1123 21:15:00.173444 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88865da-26c0-4533-81f1-f1b311a00739" containerName="registry-server" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.173480 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88865da-26c0-4533-81f1-f1b311a00739" containerName="registry-server" Nov 23 21:15:00 crc kubenswrapper[4726]: E1123 21:15:00.173526 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88865da-26c0-4533-81f1-f1b311a00739" containerName="extract-utilities" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.173538 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88865da-26c0-4533-81f1-f1b311a00739" containerName="extract-utilities" Nov 23 21:15:00 crc kubenswrapper[4726]: E1123 21:15:00.173568 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88865da-26c0-4533-81f1-f1b311a00739" containerName="extract-content" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.173580 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88865da-26c0-4533-81f1-f1b311a00739" containerName="extract-content" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.173834 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88865da-26c0-4533-81f1-f1b311a00739" containerName="registry-server" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.174712 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.202119 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.202479 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.218691 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2"] Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.219691 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50a428bd-e117-4d26-96f4-43dbd8be61e5-config-volume\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.219740 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50a428bd-e117-4d26-96f4-43dbd8be61e5-secret-volume\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.219902 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm59t\" (UniqueName: \"kubernetes.io/projected/50a428bd-e117-4d26-96f4-43dbd8be61e5-kube-api-access-cm59t\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.322156 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50a428bd-e117-4d26-96f4-43dbd8be61e5-config-volume\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.322463 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50a428bd-e117-4d26-96f4-43dbd8be61e5-secret-volume\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.322620 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm59t\" (UniqueName: \"kubernetes.io/projected/50a428bd-e117-4d26-96f4-43dbd8be61e5-kube-api-access-cm59t\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.323617 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50a428bd-e117-4d26-96f4-43dbd8be61e5-config-volume\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.338685 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50a428bd-e117-4d26-96f4-43dbd8be61e5-secret-volume\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.339199 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm59t\" (UniqueName: \"kubernetes.io/projected/50a428bd-e117-4d26-96f4-43dbd8be61e5-kube-api-access-cm59t\") pod \"collect-profiles-29398875-rtqh2\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.543708 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:00 crc kubenswrapper[4726]: I1123 21:15:00.605322 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d88865da-26c0-4533-81f1-f1b311a00739" path="/var/lib/kubelet/pods/d88865da-26c0-4533-81f1-f1b311a00739/volumes" Nov 23 21:15:01 crc kubenswrapper[4726]: I1123 21:15:01.011189 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2"] Nov 23 21:15:01 crc kubenswrapper[4726]: I1123 21:15:01.685421 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" event={"ID":"50a428bd-e117-4d26-96f4-43dbd8be61e5","Type":"ContainerStarted","Data":"437b25058249fe9af7cc66983cdb570d5bc6262b0bae5acca11ec8a9edb7b8a7"} Nov 23 21:15:01 crc kubenswrapper[4726]: I1123 21:15:01.685751 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" event={"ID":"50a428bd-e117-4d26-96f4-43dbd8be61e5","Type":"ContainerStarted","Data":"284902fff9c9a5bbaab362f824437cb3374964567b20251fd9a3d0838f3a07d6"} Nov 23 21:15:01 crc kubenswrapper[4726]: I1123 21:15:01.700848 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" podStartSLOduration=1.700828972 podStartE2EDuration="1.700828972s" podCreationTimestamp="2025-11-23 21:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:15:01.697120428 +0000 UTC m=+4009.846161404" watchObservedRunningTime="2025-11-23 21:15:01.700828972 +0000 UTC m=+4009.849869938" Nov 23 21:15:02 crc kubenswrapper[4726]: I1123 21:15:02.694910 4726 generic.go:334] "Generic (PLEG): container finished" podID="50a428bd-e117-4d26-96f4-43dbd8be61e5" containerID="437b25058249fe9af7cc66983cdb570d5bc6262b0bae5acca11ec8a9edb7b8a7" exitCode=0 Nov 23 21:15:02 crc kubenswrapper[4726]: I1123 21:15:02.694960 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" event={"ID":"50a428bd-e117-4d26-96f4-43dbd8be61e5","Type":"ContainerDied","Data":"437b25058249fe9af7cc66983cdb570d5bc6262b0bae5acca11ec8a9edb7b8a7"} Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.089353 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.201744 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm59t\" (UniqueName: \"kubernetes.io/projected/50a428bd-e117-4d26-96f4-43dbd8be61e5-kube-api-access-cm59t\") pod \"50a428bd-e117-4d26-96f4-43dbd8be61e5\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.201947 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50a428bd-e117-4d26-96f4-43dbd8be61e5-config-volume\") pod \"50a428bd-e117-4d26-96f4-43dbd8be61e5\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.202107 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50a428bd-e117-4d26-96f4-43dbd8be61e5-secret-volume\") pod \"50a428bd-e117-4d26-96f4-43dbd8be61e5\" (UID: \"50a428bd-e117-4d26-96f4-43dbd8be61e5\") " Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.202775 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50a428bd-e117-4d26-96f4-43dbd8be61e5-config-volume" (OuterVolumeSpecName: "config-volume") pod "50a428bd-e117-4d26-96f4-43dbd8be61e5" (UID: "50a428bd-e117-4d26-96f4-43dbd8be61e5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.213721 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50a428bd-e117-4d26-96f4-43dbd8be61e5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "50a428bd-e117-4d26-96f4-43dbd8be61e5" (UID: "50a428bd-e117-4d26-96f4-43dbd8be61e5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.213739 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a428bd-e117-4d26-96f4-43dbd8be61e5-kube-api-access-cm59t" (OuterVolumeSpecName: "kube-api-access-cm59t") pod "50a428bd-e117-4d26-96f4-43dbd8be61e5" (UID: "50a428bd-e117-4d26-96f4-43dbd8be61e5"). InnerVolumeSpecName "kube-api-access-cm59t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.303777 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50a428bd-e117-4d26-96f4-43dbd8be61e5-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.303809 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm59t\" (UniqueName: \"kubernetes.io/projected/50a428bd-e117-4d26-96f4-43dbd8be61e5-kube-api-access-cm59t\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.303818 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50a428bd-e117-4d26-96f4-43dbd8be61e5-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.719448 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" event={"ID":"50a428bd-e117-4d26-96f4-43dbd8be61e5","Type":"ContainerDied","Data":"284902fff9c9a5bbaab362f824437cb3374964567b20251fd9a3d0838f3a07d6"} Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.720023 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="284902fff9c9a5bbaab362f824437cb3374964567b20251fd9a3d0838f3a07d6" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.719517 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398875-rtqh2" Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.777737 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj"] Nov 23 21:15:04 crc kubenswrapper[4726]: I1123 21:15:04.785807 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398830-6pqgj"] Nov 23 21:15:05 crc kubenswrapper[4726]: I1123 21:15:05.826965 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:15:05 crc kubenswrapper[4726]: I1123 21:15:05.827329 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:15:05 crc kubenswrapper[4726]: I1123 21:15:05.879494 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:15:06 crc kubenswrapper[4726]: I1123 21:15:06.601072 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b082b91-898d-411b-beb4-7c6482f4c169" path="/var/lib/kubelet/pods/4b082b91-898d-411b-beb4-7c6482f4c169/volumes" Nov 23 21:15:06 crc kubenswrapper[4726]: I1123 21:15:06.788657 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:15:06 crc kubenswrapper[4726]: I1123 21:15:06.834284 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpsnb"] Nov 23 21:15:08 crc kubenswrapper[4726]: I1123 21:15:08.047005 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:15:08 crc kubenswrapper[4726]: I1123 21:15:08.095763 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:15:08 crc kubenswrapper[4726]: I1123 21:15:08.518809 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fpcfn"] Nov 23 21:15:08 crc kubenswrapper[4726]: I1123 21:15:08.770490 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fpsnb" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerName="registry-server" containerID="cri-o://68f10f5fc1b79051213dc60029b4fcaf18d2a161474dee448f0f6a8ddf828d74" gracePeriod=2 Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.042991 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.043040 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.779404 4726 generic.go:334] "Generic (PLEG): container finished" podID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerID="68f10f5fc1b79051213dc60029b4fcaf18d2a161474dee448f0f6a8ddf828d74" exitCode=0 Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.780006 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpsnb" event={"ID":"3af6c1aa-92a2-468a-837f-af8046c6db4f","Type":"ContainerDied","Data":"68f10f5fc1b79051213dc60029b4fcaf18d2a161474dee448f0f6a8ddf828d74"} Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.780063 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpsnb" event={"ID":"3af6c1aa-92a2-468a-837f-af8046c6db4f","Type":"ContainerDied","Data":"e8bdcc21650d64e2980e0a85114d0417cd54325dad57a4c6126324ab9f69fedf"} Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.780085 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8bdcc21650d64e2980e0a85114d0417cd54325dad57a4c6126324ab9f69fedf" Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.780152 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fpcfn" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="registry-server" containerID="cri-o://da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684" gracePeriod=2 Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.782489 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.906432 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-utilities\") pod \"3af6c1aa-92a2-468a-837f-af8046c6db4f\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.906571 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-catalog-content\") pod \"3af6c1aa-92a2-468a-837f-af8046c6db4f\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.906650 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr8s6\" (UniqueName: \"kubernetes.io/projected/3af6c1aa-92a2-468a-837f-af8046c6db4f-kube-api-access-cr8s6\") pod \"3af6c1aa-92a2-468a-837f-af8046c6db4f\" (UID: \"3af6c1aa-92a2-468a-837f-af8046c6db4f\") " Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.907392 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-utilities" (OuterVolumeSpecName: "utilities") pod "3af6c1aa-92a2-468a-837f-af8046c6db4f" (UID: "3af6c1aa-92a2-468a-837f-af8046c6db4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.914649 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3af6c1aa-92a2-468a-837f-af8046c6db4f-kube-api-access-cr8s6" (OuterVolumeSpecName: "kube-api-access-cr8s6") pod "3af6c1aa-92a2-468a-837f-af8046c6db4f" (UID: "3af6c1aa-92a2-468a-837f-af8046c6db4f"). InnerVolumeSpecName "kube-api-access-cr8s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:15:09 crc kubenswrapper[4726]: I1123 21:15:09.927646 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3af6c1aa-92a2-468a-837f-af8046c6db4f" (UID: "3af6c1aa-92a2-468a-837f-af8046c6db4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.008610 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.008630 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af6c1aa-92a2-468a-837f-af8046c6db4f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.008641 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr8s6\" (UniqueName: \"kubernetes.io/projected/3af6c1aa-92a2-468a-837f-af8046c6db4f-kube-api-access-cr8s6\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.306991 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.415772 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-catalog-content\") pod \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.416097 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhrj5\" (UniqueName: \"kubernetes.io/projected/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-kube-api-access-dhrj5\") pod \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.416348 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-utilities\") pod \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\" (UID: \"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc\") " Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.416955 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-utilities" (OuterVolumeSpecName: "utilities") pod "1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" (UID: "1d53d3e6-c36a-4298-b86d-8f6c83a64ecc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.419030 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-kube-api-access-dhrj5" (OuterVolumeSpecName: "kube-api-access-dhrj5") pod "1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" (UID: "1d53d3e6-c36a-4298-b86d-8f6c83a64ecc"). InnerVolumeSpecName "kube-api-access-dhrj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.517586 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" (UID: "1d53d3e6-c36a-4298-b86d-8f6c83a64ecc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.520746 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhrj5\" (UniqueName: \"kubernetes.io/projected/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-kube-api-access-dhrj5\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.520779 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.520789 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.790784 4726 generic.go:334] "Generic (PLEG): container finished" podID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerID="da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684" exitCode=0 Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.790882 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpcfn" event={"ID":"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc","Type":"ContainerDied","Data":"da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684"} Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.790924 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpsnb" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.790945 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fpcfn" event={"ID":"1d53d3e6-c36a-4298-b86d-8f6c83a64ecc","Type":"ContainerDied","Data":"b6e94cf90341006336a0fb51f12e1568b95f99e55569879342bc62925e56e2db"} Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.790949 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fpcfn" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.790973 4726 scope.go:117] "RemoveContainer" containerID="da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.813577 4726 scope.go:117] "RemoveContainer" containerID="6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.816977 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpsnb"] Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.833608 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpsnb"] Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.840720 4726 scope.go:117] "RemoveContainer" containerID="6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.842016 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fpcfn"] Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.849812 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fpcfn"] Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.864418 4726 scope.go:117] "RemoveContainer" containerID="da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684" Nov 23 21:15:10 crc kubenswrapper[4726]: E1123 21:15:10.865010 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684\": container with ID starting with da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684 not found: ID does not exist" containerID="da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.865121 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684"} err="failed to get container status \"da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684\": rpc error: code = NotFound desc = could not find container \"da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684\": container with ID starting with da7f69108061c1bfbdca49b835cbe8c2cc91f7aa8b181cf6c188521376a7f684 not found: ID does not exist" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.865215 4726 scope.go:117] "RemoveContainer" containerID="6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19" Nov 23 21:15:10 crc kubenswrapper[4726]: E1123 21:15:10.865573 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19\": container with ID starting with 6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19 not found: ID does not exist" containerID="6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.865615 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19"} err="failed to get container status \"6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19\": rpc error: code = NotFound desc = could not find container \"6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19\": container with ID starting with 6eff3dc45bfec1ce8c86cdaf0b296ad75f4795e3e79384f816fac9c1edf1ca19 not found: ID does not exist" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.865645 4726 scope.go:117] "RemoveContainer" containerID="6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6" Nov 23 21:15:10 crc kubenswrapper[4726]: E1123 21:15:10.866222 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6\": container with ID starting with 6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6 not found: ID does not exist" containerID="6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6" Nov 23 21:15:10 crc kubenswrapper[4726]: I1123 21:15:10.866325 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6"} err="failed to get container status \"6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6\": rpc error: code = NotFound desc = could not find container \"6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6\": container with ID starting with 6aabe224a13fa16eb6b58c3f59c0c6d98848c20780ef48f3948aca5a36d127d6 not found: ID does not exist" Nov 23 21:15:12 crc kubenswrapper[4726]: I1123 21:15:12.608384 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" path="/var/lib/kubelet/pods/1d53d3e6-c36a-4298-b86d-8f6c83a64ecc/volumes" Nov 23 21:15:12 crc kubenswrapper[4726]: I1123 21:15:12.610144 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" path="/var/lib/kubelet/pods/3af6c1aa-92a2-468a-837f-af8046c6db4f/volumes" Nov 23 21:15:21 crc kubenswrapper[4726]: I1123 21:15:21.292014 4726 scope.go:117] "RemoveContainer" containerID="ef63d815221c7aa0e2bab66001a7ebd5bc73bbc66c7bdb599a0234fff6037a2a" Nov 23 21:15:39 crc kubenswrapper[4726]: I1123 21:15:39.043107 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:15:39 crc kubenswrapper[4726]: I1123 21:15:39.043720 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.043112 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.043714 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.043755 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.044615 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7af01d69b4d8696a9058f0f4eaa0367543076fc2495dc5f83abbcf020b3a86d7"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.044663 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://7af01d69b4d8696a9058f0f4eaa0367543076fc2495dc5f83abbcf020b3a86d7" gracePeriod=600 Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.416686 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="7af01d69b4d8696a9058f0f4eaa0367543076fc2495dc5f83abbcf020b3a86d7" exitCode=0 Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.416788 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"7af01d69b4d8696a9058f0f4eaa0367543076fc2495dc5f83abbcf020b3a86d7"} Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.417561 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c"} Nov 23 21:16:09 crc kubenswrapper[4726]: I1123 21:16:09.417609 4726 scope.go:117] "RemoveContainer" containerID="0d9a453eb781475bef7dbe2273bb93bb0d9c742bf8177a0c72c9500277203fe5" Nov 23 21:17:40 crc kubenswrapper[4726]: I1123 21:17:40.296044 4726 generic.go:334] "Generic (PLEG): container finished" podID="efb73145-b382-4a15-89fe-6b74bc9d3594" containerID="4a5f659c08fe10cf885efd9a8c84644882d383875e496e0932c27be349b3a8fc" exitCode=0 Nov 23 21:17:40 crc kubenswrapper[4726]: I1123 21:17:40.296132 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"efb73145-b382-4a15-89fe-6b74bc9d3594","Type":"ContainerDied","Data":"4a5f659c08fe10cf885efd9a8c84644882d383875e496e0932c27be349b3a8fc"} Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.715511 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856299 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856352 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-config-data\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856389 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-workdir\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856554 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ssh-key\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856600 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-temporary\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856671 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config-secret\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856690 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856733 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ca-certs\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.856771 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pcxg\" (UniqueName: \"kubernetes.io/projected/efb73145-b382-4a15-89fe-6b74bc9d3594-kube-api-access-8pcxg\") pod \"efb73145-b382-4a15-89fe-6b74bc9d3594\" (UID: \"efb73145-b382-4a15-89fe-6b74bc9d3594\") " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.857508 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-config-data" (OuterVolumeSpecName: "config-data") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.858077 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.862746 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.863962 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efb73145-b382-4a15-89fe-6b74bc9d3594-kube-api-access-8pcxg" (OuterVolumeSpecName: "kube-api-access-8pcxg") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "kube-api-access-8pcxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.875127 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.885858 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.886428 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.894495 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.921743 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "efb73145-b382-4a15-89fe-6b74bc9d3594" (UID: "efb73145-b382-4a15-89fe-6b74bc9d3594"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.959103 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.959138 4726 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.959151 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.960200 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.960220 4726 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efb73145-b382-4a15-89fe-6b74bc9d3594-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.960229 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pcxg\" (UniqueName: \"kubernetes.io/projected/efb73145-b382-4a15-89fe-6b74bc9d3594-kube-api-access-8pcxg\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.960247 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.960255 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efb73145-b382-4a15-89fe-6b74bc9d3594-config-data\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.960263 4726 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efb73145-b382-4a15-89fe-6b74bc9d3594-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:41 crc kubenswrapper[4726]: I1123 21:17:41.987165 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 23 21:17:42 crc kubenswrapper[4726]: I1123 21:17:42.062505 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 23 21:17:42 crc kubenswrapper[4726]: I1123 21:17:42.318905 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"efb73145-b382-4a15-89fe-6b74bc9d3594","Type":"ContainerDied","Data":"871bee6dc4b1effc188424a1303bdab763b193bdfc728d8a7dafd29fce604fcc"} Nov 23 21:17:42 crc kubenswrapper[4726]: I1123 21:17:42.318944 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="871bee6dc4b1effc188424a1303bdab763b193bdfc728d8a7dafd29fce604fcc" Nov 23 21:17:42 crc kubenswrapper[4726]: I1123 21:17:42.319023 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.897271 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 23 21:17:45 crc kubenswrapper[4726]: E1123 21:17:45.899061 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerName="extract-content" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899086 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerName="extract-content" Nov 23 21:17:45 crc kubenswrapper[4726]: E1123 21:17:45.899109 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerName="registry-server" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899120 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerName="registry-server" Nov 23 21:17:45 crc kubenswrapper[4726]: E1123 21:17:45.899150 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="extract-content" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899162 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="extract-content" Nov 23 21:17:45 crc kubenswrapper[4726]: E1123 21:17:45.899189 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerName="extract-utilities" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899200 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerName="extract-utilities" Nov 23 21:17:45 crc kubenswrapper[4726]: E1123 21:17:45.899213 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a428bd-e117-4d26-96f4-43dbd8be61e5" containerName="collect-profiles" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899225 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a428bd-e117-4d26-96f4-43dbd8be61e5" containerName="collect-profiles" Nov 23 21:17:45 crc kubenswrapper[4726]: E1123 21:17:45.899247 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="extract-utilities" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899257 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="extract-utilities" Nov 23 21:17:45 crc kubenswrapper[4726]: E1123 21:17:45.899283 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb73145-b382-4a15-89fe-6b74bc9d3594" containerName="tempest-tests-tempest-tests-runner" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899292 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb73145-b382-4a15-89fe-6b74bc9d3594" containerName="tempest-tests-tempest-tests-runner" Nov 23 21:17:45 crc kubenswrapper[4726]: E1123 21:17:45.899319 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="registry-server" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899330 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="registry-server" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899655 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a428bd-e117-4d26-96f4-43dbd8be61e5" containerName="collect-profiles" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899691 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="efb73145-b382-4a15-89fe-6b74bc9d3594" containerName="tempest-tests-tempest-tests-runner" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899703 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d53d3e6-c36a-4298-b86d-8f6c83a64ecc" containerName="registry-server" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.899728 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3af6c1aa-92a2-468a-837f-af8046c6db4f" containerName="registry-server" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.900554 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.906449 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-xbl4l" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.906606 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.951812 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sct8g\" (UniqueName: \"kubernetes.io/projected/eeaefcf2-c7a8-4b74-b883-9b05a5e01a23-kube-api-access-sct8g\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:45 crc kubenswrapper[4726]: I1123 21:17:45.952290 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:46 crc kubenswrapper[4726]: I1123 21:17:46.055229 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sct8g\" (UniqueName: \"kubernetes.io/projected/eeaefcf2-c7a8-4b74-b883-9b05a5e01a23-kube-api-access-sct8g\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:46 crc kubenswrapper[4726]: I1123 21:17:46.055439 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:46 crc kubenswrapper[4726]: I1123 21:17:46.056265 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:46 crc kubenswrapper[4726]: I1123 21:17:46.082514 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sct8g\" (UniqueName: \"kubernetes.io/projected/eeaefcf2-c7a8-4b74-b883-9b05a5e01a23-kube-api-access-sct8g\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:46 crc kubenswrapper[4726]: I1123 21:17:46.084133 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:46 crc kubenswrapper[4726]: I1123 21:17:46.223823 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 23 21:17:46 crc kubenswrapper[4726]: I1123 21:17:46.702691 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 23 21:17:47 crc kubenswrapper[4726]: W1123 21:17:47.303675 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeeaefcf2_c7a8_4b74_b883_9b05a5e01a23.slice/crio-9b44fc88b27013e05be9a38302d4b104cdcd458acba3ad8a5bb4138552a3735c WatchSource:0}: Error finding container 9b44fc88b27013e05be9a38302d4b104cdcd458acba3ad8a5bb4138552a3735c: Status 404 returned error can't find the container with id 9b44fc88b27013e05be9a38302d4b104cdcd458acba3ad8a5bb4138552a3735c Nov 23 21:17:47 crc kubenswrapper[4726]: I1123 21:17:47.310264 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 21:17:47 crc kubenswrapper[4726]: I1123 21:17:47.371060 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23","Type":"ContainerStarted","Data":"9b44fc88b27013e05be9a38302d4b104cdcd458acba3ad8a5bb4138552a3735c"} Nov 23 21:17:49 crc kubenswrapper[4726]: I1123 21:17:49.461332 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"eeaefcf2-c7a8-4b74-b883-9b05a5e01a23","Type":"ContainerStarted","Data":"93813499ddd7fcfcacda8924825c13bbf33addb1fe4400fc9f47d1b08f574f8f"} Nov 23 21:17:49 crc kubenswrapper[4726]: I1123 21:17:49.480589 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.483818619 podStartE2EDuration="4.480534811s" podCreationTimestamp="2025-11-23 21:17:45 +0000 UTC" firstStartedPulling="2025-11-23 21:17:47.310046421 +0000 UTC m=+4175.459087377" lastFinishedPulling="2025-11-23 21:17:48.306762603 +0000 UTC m=+4176.455803569" observedRunningTime="2025-11-23 21:17:49.473116372 +0000 UTC m=+4177.622157348" watchObservedRunningTime="2025-11-23 21:17:49.480534811 +0000 UTC m=+4177.629575777" Nov 23 21:18:09 crc kubenswrapper[4726]: I1123 21:18:09.043618 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:18:09 crc kubenswrapper[4726]: I1123 21:18:09.044183 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.585934 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gz4k6/must-gather-cw2lq"] Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.587772 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.598406 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gz4k6"/"openshift-service-ca.crt" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.598474 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gz4k6"/"default-dockercfg-7mqqn" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.598628 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gz4k6"/"kube-root-ca.crt" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.623793 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gz4k6/must-gather-cw2lq"] Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.673152 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f56pq\" (UniqueName: \"kubernetes.io/projected/d5709a4c-d883-4f50-b622-f482770107ff-kube-api-access-f56pq\") pod \"must-gather-cw2lq\" (UID: \"d5709a4c-d883-4f50-b622-f482770107ff\") " pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.673523 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d5709a4c-d883-4f50-b622-f482770107ff-must-gather-output\") pod \"must-gather-cw2lq\" (UID: \"d5709a4c-d883-4f50-b622-f482770107ff\") " pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.775642 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f56pq\" (UniqueName: \"kubernetes.io/projected/d5709a4c-d883-4f50-b622-f482770107ff-kube-api-access-f56pq\") pod \"must-gather-cw2lq\" (UID: \"d5709a4c-d883-4f50-b622-f482770107ff\") " pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.776480 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d5709a4c-d883-4f50-b622-f482770107ff-must-gather-output\") pod \"must-gather-cw2lq\" (UID: \"d5709a4c-d883-4f50-b622-f482770107ff\") " pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.776073 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d5709a4c-d883-4f50-b622-f482770107ff-must-gather-output\") pod \"must-gather-cw2lq\" (UID: \"d5709a4c-d883-4f50-b622-f482770107ff\") " pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.796955 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f56pq\" (UniqueName: \"kubernetes.io/projected/d5709a4c-d883-4f50-b622-f482770107ff-kube-api-access-f56pq\") pod \"must-gather-cw2lq\" (UID: \"d5709a4c-d883-4f50-b622-f482770107ff\") " pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:18:12 crc kubenswrapper[4726]: I1123 21:18:12.908545 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:18:13 crc kubenswrapper[4726]: I1123 21:18:13.379794 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gz4k6/must-gather-cw2lq"] Nov 23 21:18:13 crc kubenswrapper[4726]: W1123 21:18:13.495839 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5709a4c_d883_4f50_b622_f482770107ff.slice/crio-b503f638c4b07c075c20bb9daea74bc428c3bf4be161ba6b703c07eb22f17557 WatchSource:0}: Error finding container b503f638c4b07c075c20bb9daea74bc428c3bf4be161ba6b703c07eb22f17557: Status 404 returned error can't find the container with id b503f638c4b07c075c20bb9daea74bc428c3bf4be161ba6b703c07eb22f17557 Nov 23 21:18:13 crc kubenswrapper[4726]: I1123 21:18:13.735826 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" event={"ID":"d5709a4c-d883-4f50-b622-f482770107ff","Type":"ContainerStarted","Data":"b503f638c4b07c075c20bb9daea74bc428c3bf4be161ba6b703c07eb22f17557"} Nov 23 21:18:21 crc kubenswrapper[4726]: I1123 21:18:21.819630 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" event={"ID":"d5709a4c-d883-4f50-b622-f482770107ff","Type":"ContainerStarted","Data":"aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070"} Nov 23 21:18:21 crc kubenswrapper[4726]: I1123 21:18:21.820327 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" event={"ID":"d5709a4c-d883-4f50-b622-f482770107ff","Type":"ContainerStarted","Data":"4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce"} Nov 23 21:18:21 crc kubenswrapper[4726]: I1123 21:18:21.844402 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" podStartSLOduration=2.16273781 podStartE2EDuration="9.844381728s" podCreationTimestamp="2025-11-23 21:18:12 +0000 UTC" firstStartedPulling="2025-11-23 21:18:13.497818069 +0000 UTC m=+4201.646859025" lastFinishedPulling="2025-11-23 21:18:21.179461987 +0000 UTC m=+4209.328502943" observedRunningTime="2025-11-23 21:18:21.835612241 +0000 UTC m=+4209.984653217" watchObservedRunningTime="2025-11-23 21:18:21.844381728 +0000 UTC m=+4209.993422684" Nov 23 21:18:27 crc kubenswrapper[4726]: I1123 21:18:27.998166 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-rzhnv"] Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.000680 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.073368 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65a28240-dc70-4ebc-93e1-166ec4933664-host\") pod \"crc-debug-rzhnv\" (UID: \"65a28240-dc70-4ebc-93e1-166ec4933664\") " pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.073604 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jzv7\" (UniqueName: \"kubernetes.io/projected/65a28240-dc70-4ebc-93e1-166ec4933664-kube-api-access-5jzv7\") pod \"crc-debug-rzhnv\" (UID: \"65a28240-dc70-4ebc-93e1-166ec4933664\") " pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.174956 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65a28240-dc70-4ebc-93e1-166ec4933664-host\") pod \"crc-debug-rzhnv\" (UID: \"65a28240-dc70-4ebc-93e1-166ec4933664\") " pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.175048 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65a28240-dc70-4ebc-93e1-166ec4933664-host\") pod \"crc-debug-rzhnv\" (UID: \"65a28240-dc70-4ebc-93e1-166ec4933664\") " pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.175237 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jzv7\" (UniqueName: \"kubernetes.io/projected/65a28240-dc70-4ebc-93e1-166ec4933664-kube-api-access-5jzv7\") pod \"crc-debug-rzhnv\" (UID: \"65a28240-dc70-4ebc-93e1-166ec4933664\") " pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.194251 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jzv7\" (UniqueName: \"kubernetes.io/projected/65a28240-dc70-4ebc-93e1-166ec4933664-kube-api-access-5jzv7\") pod \"crc-debug-rzhnv\" (UID: \"65a28240-dc70-4ebc-93e1-166ec4933664\") " pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.331111 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:18:28 crc kubenswrapper[4726]: I1123 21:18:28.903408 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" event={"ID":"65a28240-dc70-4ebc-93e1-166ec4933664","Type":"ContainerStarted","Data":"b812bddbd1ae1258f71fd227915b10a2f82ffa5b7d65c5dedf04716c5bdbe6dd"} Nov 23 21:18:39 crc kubenswrapper[4726]: I1123 21:18:39.043207 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:18:39 crc kubenswrapper[4726]: I1123 21:18:39.043670 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:18:40 crc kubenswrapper[4726]: I1123 21:18:40.007769 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" event={"ID":"65a28240-dc70-4ebc-93e1-166ec4933664","Type":"ContainerStarted","Data":"c4ca43741fb2457afdd49fd87e81e61ebc259bf8d34cf1c9f0e0d986ecc656a1"} Nov 23 21:18:40 crc kubenswrapper[4726]: I1123 21:18:40.023166 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" podStartSLOduration=1.960726254 podStartE2EDuration="13.023146285s" podCreationTimestamp="2025-11-23 21:18:27 +0000 UTC" firstStartedPulling="2025-11-23 21:18:28.378048133 +0000 UTC m=+4216.527089089" lastFinishedPulling="2025-11-23 21:18:39.440468164 +0000 UTC m=+4227.589509120" observedRunningTime="2025-11-23 21:18:40.019191454 +0000 UTC m=+4228.168232410" watchObservedRunningTime="2025-11-23 21:18:40.023146285 +0000 UTC m=+4228.172187241" Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.042715 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.043390 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.043463 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.044324 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.044382 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" gracePeriod=600 Nov 23 21:19:09 crc kubenswrapper[4726]: E1123 21:19:09.181290 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.275899 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" exitCode=0 Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.276237 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c"} Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.276272 4726 scope.go:117] "RemoveContainer" containerID="7af01d69b4d8696a9058f0f4eaa0367543076fc2495dc5f83abbcf020b3a86d7" Nov 23 21:19:09 crc kubenswrapper[4726]: I1123 21:19:09.277186 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:19:09 crc kubenswrapper[4726]: E1123 21:19:09.277537 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:19:20 crc kubenswrapper[4726]: I1123 21:19:20.589489 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:19:20 crc kubenswrapper[4726]: E1123 21:19:20.590345 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:19:23 crc kubenswrapper[4726]: I1123 21:19:23.394664 4726 generic.go:334] "Generic (PLEG): container finished" podID="65a28240-dc70-4ebc-93e1-166ec4933664" containerID="c4ca43741fb2457afdd49fd87e81e61ebc259bf8d34cf1c9f0e0d986ecc656a1" exitCode=0 Nov 23 21:19:23 crc kubenswrapper[4726]: I1123 21:19:23.394751 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" event={"ID":"65a28240-dc70-4ebc-93e1-166ec4933664","Type":"ContainerDied","Data":"c4ca43741fb2457afdd49fd87e81e61ebc259bf8d34cf1c9f0e0d986ecc656a1"} Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.540160 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.566333 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-rzhnv"] Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.574604 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-rzhnv"] Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.630177 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65a28240-dc70-4ebc-93e1-166ec4933664-host\") pod \"65a28240-dc70-4ebc-93e1-166ec4933664\" (UID: \"65a28240-dc70-4ebc-93e1-166ec4933664\") " Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.630319 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65a28240-dc70-4ebc-93e1-166ec4933664-host" (OuterVolumeSpecName: "host") pod "65a28240-dc70-4ebc-93e1-166ec4933664" (UID: "65a28240-dc70-4ebc-93e1-166ec4933664"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.630345 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jzv7\" (UniqueName: \"kubernetes.io/projected/65a28240-dc70-4ebc-93e1-166ec4933664-kube-api-access-5jzv7\") pod \"65a28240-dc70-4ebc-93e1-166ec4933664\" (UID: \"65a28240-dc70-4ebc-93e1-166ec4933664\") " Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.630857 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65a28240-dc70-4ebc-93e1-166ec4933664-host\") on node \"crc\" DevicePath \"\"" Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.636237 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65a28240-dc70-4ebc-93e1-166ec4933664-kube-api-access-5jzv7" (OuterVolumeSpecName: "kube-api-access-5jzv7") pod "65a28240-dc70-4ebc-93e1-166ec4933664" (UID: "65a28240-dc70-4ebc-93e1-166ec4933664"). InnerVolumeSpecName "kube-api-access-5jzv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:19:24 crc kubenswrapper[4726]: I1123 21:19:24.733075 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jzv7\" (UniqueName: \"kubernetes.io/projected/65a28240-dc70-4ebc-93e1-166ec4933664-kube-api-access-5jzv7\") on node \"crc\" DevicePath \"\"" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.424052 4726 scope.go:117] "RemoveContainer" containerID="c4ca43741fb2457afdd49fd87e81e61ebc259bf8d34cf1c9f0e0d986ecc656a1" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.424113 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-rzhnv" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.752520 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-smgdh"] Nov 23 21:19:25 crc kubenswrapper[4726]: E1123 21:19:25.752937 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a28240-dc70-4ebc-93e1-166ec4933664" containerName="container-00" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.752953 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a28240-dc70-4ebc-93e1-166ec4933664" containerName="container-00" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.753204 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="65a28240-dc70-4ebc-93e1-166ec4933664" containerName="container-00" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.753932 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.854088 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c99680ca-391d-479b-97b7-e6e55f87d21b-host\") pod \"crc-debug-smgdh\" (UID: \"c99680ca-391d-479b-97b7-e6e55f87d21b\") " pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.854168 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksrq4\" (UniqueName: \"kubernetes.io/projected/c99680ca-391d-479b-97b7-e6e55f87d21b-kube-api-access-ksrq4\") pod \"crc-debug-smgdh\" (UID: \"c99680ca-391d-479b-97b7-e6e55f87d21b\") " pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.956090 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c99680ca-391d-479b-97b7-e6e55f87d21b-host\") pod \"crc-debug-smgdh\" (UID: \"c99680ca-391d-479b-97b7-e6e55f87d21b\") " pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.956369 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksrq4\" (UniqueName: \"kubernetes.io/projected/c99680ca-391d-479b-97b7-e6e55f87d21b-kube-api-access-ksrq4\") pod \"crc-debug-smgdh\" (UID: \"c99680ca-391d-479b-97b7-e6e55f87d21b\") " pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.956257 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c99680ca-391d-479b-97b7-e6e55f87d21b-host\") pod \"crc-debug-smgdh\" (UID: \"c99680ca-391d-479b-97b7-e6e55f87d21b\") " pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:25 crc kubenswrapper[4726]: I1123 21:19:25.981416 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksrq4\" (UniqueName: \"kubernetes.io/projected/c99680ca-391d-479b-97b7-e6e55f87d21b-kube-api-access-ksrq4\") pod \"crc-debug-smgdh\" (UID: \"c99680ca-391d-479b-97b7-e6e55f87d21b\") " pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:26 crc kubenswrapper[4726]: I1123 21:19:26.068383 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:26 crc kubenswrapper[4726]: I1123 21:19:26.435386 4726 generic.go:334] "Generic (PLEG): container finished" podID="c99680ca-391d-479b-97b7-e6e55f87d21b" containerID="81d5f91df0f17432bdea6edbba2d11a0ad818aa014306a3f300498a14633e360" exitCode=0 Nov 23 21:19:26 crc kubenswrapper[4726]: I1123 21:19:26.435434 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/crc-debug-smgdh" event={"ID":"c99680ca-391d-479b-97b7-e6e55f87d21b","Type":"ContainerDied","Data":"81d5f91df0f17432bdea6edbba2d11a0ad818aa014306a3f300498a14633e360"} Nov 23 21:19:26 crc kubenswrapper[4726]: I1123 21:19:26.435725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/crc-debug-smgdh" event={"ID":"c99680ca-391d-479b-97b7-e6e55f87d21b","Type":"ContainerStarted","Data":"15b90686cb0991839451929c06c5f79a2c71c425b6cbe6fdc7602cfe2b0df96b"} Nov 23 21:19:26 crc kubenswrapper[4726]: I1123 21:19:26.599969 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65a28240-dc70-4ebc-93e1-166ec4933664" path="/var/lib/kubelet/pods/65a28240-dc70-4ebc-93e1-166ec4933664/volumes" Nov 23 21:19:26 crc kubenswrapper[4726]: I1123 21:19:26.850466 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-smgdh"] Nov 23 21:19:26 crc kubenswrapper[4726]: I1123 21:19:26.860998 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-smgdh"] Nov 23 21:19:27 crc kubenswrapper[4726]: I1123 21:19:27.535348 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:27 crc kubenswrapper[4726]: I1123 21:19:27.687379 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c99680ca-391d-479b-97b7-e6e55f87d21b-host\") pod \"c99680ca-391d-479b-97b7-e6e55f87d21b\" (UID: \"c99680ca-391d-479b-97b7-e6e55f87d21b\") " Nov 23 21:19:27 crc kubenswrapper[4726]: I1123 21:19:27.687459 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksrq4\" (UniqueName: \"kubernetes.io/projected/c99680ca-391d-479b-97b7-e6e55f87d21b-kube-api-access-ksrq4\") pod \"c99680ca-391d-479b-97b7-e6e55f87d21b\" (UID: \"c99680ca-391d-479b-97b7-e6e55f87d21b\") " Nov 23 21:19:27 crc kubenswrapper[4726]: I1123 21:19:27.687495 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c99680ca-391d-479b-97b7-e6e55f87d21b-host" (OuterVolumeSpecName: "host") pod "c99680ca-391d-479b-97b7-e6e55f87d21b" (UID: "c99680ca-391d-479b-97b7-e6e55f87d21b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:19:27 crc kubenswrapper[4726]: I1123 21:19:27.687999 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c99680ca-391d-479b-97b7-e6e55f87d21b-host\") on node \"crc\" DevicePath \"\"" Nov 23 21:19:27 crc kubenswrapper[4726]: I1123 21:19:27.708223 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c99680ca-391d-479b-97b7-e6e55f87d21b-kube-api-access-ksrq4" (OuterVolumeSpecName: "kube-api-access-ksrq4") pod "c99680ca-391d-479b-97b7-e6e55f87d21b" (UID: "c99680ca-391d-479b-97b7-e6e55f87d21b"). InnerVolumeSpecName "kube-api-access-ksrq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:19:27 crc kubenswrapper[4726]: I1123 21:19:27.789881 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksrq4\" (UniqueName: \"kubernetes.io/projected/c99680ca-391d-479b-97b7-e6e55f87d21b-kube-api-access-ksrq4\") on node \"crc\" DevicePath \"\"" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.043435 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-pq56z"] Nov 23 21:19:28 crc kubenswrapper[4726]: E1123 21:19:28.044345 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c99680ca-391d-479b-97b7-e6e55f87d21b" containerName="container-00" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.044418 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c99680ca-391d-479b-97b7-e6e55f87d21b" containerName="container-00" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.044684 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c99680ca-391d-479b-97b7-e6e55f87d21b" containerName="container-00" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.045342 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.196165 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-host\") pod \"crc-debug-pq56z\" (UID: \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\") " pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.196252 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpcqv\" (UniqueName: \"kubernetes.io/projected/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-kube-api-access-mpcqv\") pod \"crc-debug-pq56z\" (UID: \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\") " pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.298064 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-host\") pod \"crc-debug-pq56z\" (UID: \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\") " pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.298139 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpcqv\" (UniqueName: \"kubernetes.io/projected/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-kube-api-access-mpcqv\") pod \"crc-debug-pq56z\" (UID: \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\") " pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.298539 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-host\") pod \"crc-debug-pq56z\" (UID: \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\") " pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.315326 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpcqv\" (UniqueName: \"kubernetes.io/projected/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-kube-api-access-mpcqv\") pod \"crc-debug-pq56z\" (UID: \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\") " pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.360975 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:28 crc kubenswrapper[4726]: W1123 21:19:28.383485 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc594fef9_43cd_4693_9f79_6ab4d8a8b5fd.slice/crio-2fd27bc0cf9c3fee130a65ddff1202aecb8b216301f71b3ed6be57cddf410a89 WatchSource:0}: Error finding container 2fd27bc0cf9c3fee130a65ddff1202aecb8b216301f71b3ed6be57cddf410a89: Status 404 returned error can't find the container with id 2fd27bc0cf9c3fee130a65ddff1202aecb8b216301f71b3ed6be57cddf410a89 Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.454052 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15b90686cb0991839451929c06c5f79a2c71c425b6cbe6fdc7602cfe2b0df96b" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.454129 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-smgdh" Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.457105 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/crc-debug-pq56z" event={"ID":"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd","Type":"ContainerStarted","Data":"2fd27bc0cf9c3fee130a65ddff1202aecb8b216301f71b3ed6be57cddf410a89"} Nov 23 21:19:28 crc kubenswrapper[4726]: I1123 21:19:28.601708 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c99680ca-391d-479b-97b7-e6e55f87d21b" path="/var/lib/kubelet/pods/c99680ca-391d-479b-97b7-e6e55f87d21b/volumes" Nov 23 21:19:29 crc kubenswrapper[4726]: I1123 21:19:29.474027 4726 generic.go:334] "Generic (PLEG): container finished" podID="c594fef9-43cd-4693-9f79-6ab4d8a8b5fd" containerID="2ee121ddc816e898c5f61ab80a2464216a2d56d7b039a9585f4ed2306b875a9a" exitCode=0 Nov 23 21:19:29 crc kubenswrapper[4726]: I1123 21:19:29.474205 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/crc-debug-pq56z" event={"ID":"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd","Type":"ContainerDied","Data":"2ee121ddc816e898c5f61ab80a2464216a2d56d7b039a9585f4ed2306b875a9a"} Nov 23 21:19:29 crc kubenswrapper[4726]: I1123 21:19:29.534729 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-pq56z"] Nov 23 21:19:29 crc kubenswrapper[4726]: I1123 21:19:29.552590 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gz4k6/crc-debug-pq56z"] Nov 23 21:19:30 crc kubenswrapper[4726]: I1123 21:19:30.657427 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:30 crc kubenswrapper[4726]: I1123 21:19:30.751821 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpcqv\" (UniqueName: \"kubernetes.io/projected/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-kube-api-access-mpcqv\") pod \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\" (UID: \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\") " Nov 23 21:19:30 crc kubenswrapper[4726]: I1123 21:19:30.752000 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-host\") pod \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\" (UID: \"c594fef9-43cd-4693-9f79-6ab4d8a8b5fd\") " Nov 23 21:19:30 crc kubenswrapper[4726]: I1123 21:19:30.753137 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-host" (OuterVolumeSpecName: "host") pod "c594fef9-43cd-4693-9f79-6ab4d8a8b5fd" (UID: "c594fef9-43cd-4693-9f79-6ab4d8a8b5fd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:19:30 crc kubenswrapper[4726]: I1123 21:19:30.758841 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-kube-api-access-mpcqv" (OuterVolumeSpecName: "kube-api-access-mpcqv") pod "c594fef9-43cd-4693-9f79-6ab4d8a8b5fd" (UID: "c594fef9-43cd-4693-9f79-6ab4d8a8b5fd"). InnerVolumeSpecName "kube-api-access-mpcqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:19:30 crc kubenswrapper[4726]: I1123 21:19:30.854537 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpcqv\" (UniqueName: \"kubernetes.io/projected/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-kube-api-access-mpcqv\") on node \"crc\" DevicePath \"\"" Nov 23 21:19:30 crc kubenswrapper[4726]: I1123 21:19:30.854570 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd-host\") on node \"crc\" DevicePath \"\"" Nov 23 21:19:31 crc kubenswrapper[4726]: I1123 21:19:31.517504 4726 scope.go:117] "RemoveContainer" containerID="2ee121ddc816e898c5f61ab80a2464216a2d56d7b039a9585f4ed2306b875a9a" Nov 23 21:19:31 crc kubenswrapper[4726]: I1123 21:19:31.517615 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/crc-debug-pq56z" Nov 23 21:19:32 crc kubenswrapper[4726]: I1123 21:19:32.604835 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c594fef9-43cd-4693-9f79-6ab4d8a8b5fd" path="/var/lib/kubelet/pods/c594fef9-43cd-4693-9f79-6ab4d8a8b5fd/volumes" Nov 23 21:19:34 crc kubenswrapper[4726]: I1123 21:19:34.593901 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:19:34 crc kubenswrapper[4726]: E1123 21:19:34.594420 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:19:47 crc kubenswrapper[4726]: I1123 21:19:47.590224 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:19:47 crc kubenswrapper[4726]: E1123 21:19:47.591037 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:20:02 crc kubenswrapper[4726]: I1123 21:20:02.604656 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:20:02 crc kubenswrapper[4726]: E1123 21:20:02.605644 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:20:15 crc kubenswrapper[4726]: I1123 21:20:15.589414 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:20:15 crc kubenswrapper[4726]: E1123 21:20:15.590394 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:20:27 crc kubenswrapper[4726]: I1123 21:20:27.128897 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-85855866bd-77958_cddea6c0-9834-4a47-b30c-ef32194ef6ea/barbican-api/0.log" Nov 23 21:20:27 crc kubenswrapper[4726]: I1123 21:20:27.589468 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:20:27 crc kubenswrapper[4726]: E1123 21:20:27.591186 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:20:27 crc kubenswrapper[4726]: I1123 21:20:27.831336 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-85855866bd-77958_cddea6c0-9834-4a47-b30c-ef32194ef6ea/barbican-api-log/0.log" Nov 23 21:20:27 crc kubenswrapper[4726]: I1123 21:20:27.872353 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-644b66b48-csfsq_02190dad-4d68-4cc1-937c-a45b260b4a95/barbican-keystone-listener-log/0.log" Nov 23 21:20:27 crc kubenswrapper[4726]: I1123 21:20:27.892520 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-644b66b48-csfsq_02190dad-4d68-4cc1-937c-a45b260b4a95/barbican-keystone-listener/0.log" Nov 23 21:20:28 crc kubenswrapper[4726]: I1123 21:20:28.072532 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b6bd669d5-fsxlx_3041645e-f5e5-4faf-b2de-365bcd156177/barbican-worker-log/0.log" Nov 23 21:20:28 crc kubenswrapper[4726]: I1123 21:20:28.132141 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b6bd669d5-fsxlx_3041645e-f5e5-4faf-b2de-365bcd156177/barbican-worker/0.log" Nov 23 21:20:28 crc kubenswrapper[4726]: I1123 21:20:28.316306 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7_e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:28 crc kubenswrapper[4726]: I1123 21:20:28.359201 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a3f087de-9790-4860-b3bd-1ab9e677d4bd/ceilometer-central-agent/0.log" Nov 23 21:20:28 crc kubenswrapper[4726]: I1123 21:20:28.436288 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a3f087de-9790-4860-b3bd-1ab9e677d4bd/ceilometer-notification-agent/0.log" Nov 23 21:20:28 crc kubenswrapper[4726]: I1123 21:20:28.529209 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a3f087de-9790-4860-b3bd-1ab9e677d4bd/proxy-httpd/0.log" Nov 23 21:20:28 crc kubenswrapper[4726]: I1123 21:20:28.583859 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a3f087de-9790-4860-b3bd-1ab9e677d4bd/sg-core/0.log" Nov 23 21:20:28 crc kubenswrapper[4726]: I1123 21:20:28.660504 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h_00e5cdd7-83a9-468b-a56f-db77fe0da375/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:29 crc kubenswrapper[4726]: I1123 21:20:29.266040 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2_0c1ed863-f501-4aec-bafe-2d7d828b8d78/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:29 crc kubenswrapper[4726]: I1123 21:20:29.325534 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275/cinder-api-log/0.log" Nov 23 21:20:29 crc kubenswrapper[4726]: I1123 21:20:29.413965 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275/cinder-api/0.log" Nov 23 21:20:29 crc kubenswrapper[4726]: I1123 21:20:29.566332 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047/probe/0.log" Nov 23 21:20:29 crc kubenswrapper[4726]: I1123 21:20:29.678709 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047/cinder-backup/0.log" Nov 23 21:20:29 crc kubenswrapper[4726]: I1123 21:20:29.871914 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5462b4dd-3aaf-4511-b08f-4cfe23a07a0d/cinder-scheduler/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.003997 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5462b4dd-3aaf-4511-b08f-4cfe23a07a0d/probe/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.062595 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a937c3c1-2bbf-4cf8-b5ac-9b55755df0af/cinder-volume/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.177228 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a937c3c1-2bbf-4cf8-b5ac-9b55755df0af/probe/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.322078 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-rd42k_74d7e759-c823-49e1-9480-87d93552cd4c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.407015 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr_77123683-ed45-435f-b733-f5cc3822169c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.610039 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7ddf789dd7-wznxr_6c62393d-31e9-4a24-9976-ffcf175d1443/init/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.746750 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7ddf789dd7-wznxr_6c62393d-31e9-4a24-9976-ffcf175d1443/init/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.851408 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb/glance-httpd/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.868539 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb/glance-log/0.log" Nov 23 21:20:30 crc kubenswrapper[4726]: I1123 21:20:30.930588 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7ddf789dd7-wznxr_6c62393d-31e9-4a24-9976-ffcf175d1443/dnsmasq-dns/0.log" Nov 23 21:20:31 crc kubenswrapper[4726]: I1123 21:20:31.054658 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_97a134ee-4835-43e2-82b3-e1f19fa691cc/glance-httpd/0.log" Nov 23 21:20:31 crc kubenswrapper[4726]: I1123 21:20:31.121556 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_97a134ee-4835-43e2-82b3-e1f19fa691cc/glance-log/0.log" Nov 23 21:20:31 crc kubenswrapper[4726]: I1123 21:20:31.251712 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c65c799bb-97hbb_30f69dae-368c-4ccf-9d4b-e7f943d7e662/horizon/0.log" Nov 23 21:20:31 crc kubenswrapper[4726]: I1123 21:20:31.418120 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c65c799bb-97hbb_30f69dae-368c-4ccf-9d4b-e7f943d7e662/horizon-log/0.log" Nov 23 21:20:31 crc kubenswrapper[4726]: I1123 21:20:31.463446 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d_01fcd196-1da1-4fb2-a900-f3960109b195/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:31 crc kubenswrapper[4726]: I1123 21:20:31.523837 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-972gs_541068f2-1c18-4bc7-8c86-51f768e97783/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:31 crc kubenswrapper[4726]: I1123 21:20:31.724204 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29398861-kdlcf_d2340c37-0f69-48f7-94c1-e417bb7844a5/keystone-cron/0.log" Nov 23 21:20:31 crc kubenswrapper[4726]: I1123 21:20:31.783948 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-b76988558-2489x_1071728a-1612-4dc4-9cbe-25c043998eb5/keystone-api/0.log" Nov 23 21:20:32 crc kubenswrapper[4726]: I1123 21:20:32.134213 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_71a1f62d-5ad3-450a-824a-bff237a8252e/kube-state-metrics/0.log" Nov 23 21:20:32 crc kubenswrapper[4726]: I1123 21:20:32.270061 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f82bda1a-eb9c-4e50-9700-85c8f6a55336/manila-api/0.log" Nov 23 21:20:32 crc kubenswrapper[4726]: I1123 21:20:32.314363 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n_4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:32 crc kubenswrapper[4726]: I1123 21:20:32.389422 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f82bda1a-eb9c-4e50-9700-85c8f6a55336/manila-api-log/0.log" Nov 23 21:20:32 crc kubenswrapper[4726]: I1123 21:20:32.599733 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_78ba73f5-3651-4b4b-9f8c-58fb451a655f/probe/0.log" Nov 23 21:20:32 crc kubenswrapper[4726]: I1123 21:20:32.610976 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_78ba73f5-3651-4b4b-9f8c-58fb451a655f/manila-scheduler/0.log" Nov 23 21:20:32 crc kubenswrapper[4726]: I1123 21:20:32.711686 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_ccbab0a3-e42f-42fd-8adb-ed1bae00642c/manila-share/0.log" Nov 23 21:20:32 crc kubenswrapper[4726]: I1123 21:20:32.864249 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_ccbab0a3-e42f-42fd-8adb-ed1bae00642c/probe/0.log" Nov 23 21:20:33 crc kubenswrapper[4726]: I1123 21:20:33.070684 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57cf6d944c-r5qgs_6e680832-25f3-4328-bf10-35bd83ab52b9/neutron-api/0.log" Nov 23 21:20:33 crc kubenswrapper[4726]: I1123 21:20:33.162738 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4_d233adb5-8473-4545-b91a-56ec95775b67/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:33 crc kubenswrapper[4726]: I1123 21:20:33.265459 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57cf6d944c-r5qgs_6e680832-25f3-4328-bf10-35bd83ab52b9/neutron-httpd/0.log" Nov 23 21:20:33 crc kubenswrapper[4726]: I1123 21:20:33.773917 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a327e21d-682a-4006-9ef1-b6a07276fa3d/nova-cell0-conductor-conductor/0.log" Nov 23 21:20:33 crc kubenswrapper[4726]: I1123 21:20:33.880746 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_29188283-47f2-45d2-b6ab-d3cc9b97faff/nova-api-log/0.log" Nov 23 21:20:34 crc kubenswrapper[4726]: I1123 21:20:34.123537 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_5d687a16-417f-4c97-b745-beba9b9429ae/nova-cell1-conductor-conductor/0.log" Nov 23 21:20:34 crc kubenswrapper[4726]: I1123 21:20:34.213826 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_609ae5f6-90fc-432a-820a-ca996f78c1c7/nova-cell1-novncproxy-novncproxy/0.log" Nov 23 21:20:34 crc kubenswrapper[4726]: I1123 21:20:34.231133 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_29188283-47f2-45d2-b6ab-d3cc9b97faff/nova-api-api/0.log" Nov 23 21:20:34 crc kubenswrapper[4726]: I1123 21:20:34.485412 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj_89cb9922-a339-44bc-b1be-e7ffa127cf02/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:34 crc kubenswrapper[4726]: I1123 21:20:34.516845 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7e68d296-e4d0-4462-9b15-1700400d38fa/nova-metadata-log/0.log" Nov 23 21:20:34 crc kubenswrapper[4726]: I1123 21:20:34.917536 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e324e799-3ac3-4e70-a035-822dff0d5972/nova-scheduler-scheduler/0.log" Nov 23 21:20:35 crc kubenswrapper[4726]: I1123 21:20:35.008145 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e88e0fb1-897d-454c-9a23-8187592296e5/mysql-bootstrap/0.log" Nov 23 21:20:35 crc kubenswrapper[4726]: I1123 21:20:35.242731 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e88e0fb1-897d-454c-9a23-8187592296e5/galera/0.log" Nov 23 21:20:35 crc kubenswrapper[4726]: I1123 21:20:35.295248 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e88e0fb1-897d-454c-9a23-8187592296e5/mysql-bootstrap/0.log" Nov 23 21:20:35 crc kubenswrapper[4726]: I1123 21:20:35.497271 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3647d686-8ad6-4bad-9ca6-ffbaba9bba97/mysql-bootstrap/0.log" Nov 23 21:20:35 crc kubenswrapper[4726]: I1123 21:20:35.713730 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3647d686-8ad6-4bad-9ca6-ffbaba9bba97/mysql-bootstrap/0.log" Nov 23 21:20:35 crc kubenswrapper[4726]: I1123 21:20:35.770072 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3647d686-8ad6-4bad-9ca6-ffbaba9bba97/galera/0.log" Nov 23 21:20:36 crc kubenswrapper[4726]: I1123 21:20:36.242932 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7e68d296-e4d0-4462-9b15-1700400d38fa/nova-metadata-metadata/0.log" Nov 23 21:20:36 crc kubenswrapper[4726]: I1123 21:20:36.366113 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_bfa50ecd-0a29-44c8-b219-d2e5df3709d1/openstackclient/0.log" Nov 23 21:20:36 crc kubenswrapper[4726]: I1123 21:20:36.393992 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-657sc_4cb17e51-7e2f-4012-96fd-5f6bd424429c/ovn-controller/0.log" Nov 23 21:20:36 crc kubenswrapper[4726]: I1123 21:20:36.645340 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jnv7b_fbe231ca-6343-4afb-b11b-a74c26a01a2f/openstack-network-exporter/0.log" Nov 23 21:20:36 crc kubenswrapper[4726]: I1123 21:20:36.803709 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sq7pg_ac53da75-5263-4d97-aa49-b4e8b345c854/ovsdb-server-init/0.log" Nov 23 21:20:37 crc kubenswrapper[4726]: I1123 21:20:37.005224 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sq7pg_ac53da75-5263-4d97-aa49-b4e8b345c854/ovsdb-server-init/0.log" Nov 23 21:20:37 crc kubenswrapper[4726]: I1123 21:20:37.056388 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sq7pg_ac53da75-5263-4d97-aa49-b4e8b345c854/ovs-vswitchd/0.log" Nov 23 21:20:37 crc kubenswrapper[4726]: I1123 21:20:37.067223 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sq7pg_ac53da75-5263-4d97-aa49-b4e8b345c854/ovsdb-server/0.log" Nov 23 21:20:37 crc kubenswrapper[4726]: I1123 21:20:37.268686 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d5bc7b6-0493-40ff-808b-6d13bb31c8c0/openstack-network-exporter/0.log" Nov 23 21:20:37 crc kubenswrapper[4726]: I1123 21:20:37.270733 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-q6k62_09d671dd-9524-4d36-8db8-a178093fdc70/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:37 crc kubenswrapper[4726]: I1123 21:20:37.953099 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d5bc7b6-0493-40ff-808b-6d13bb31c8c0/ovn-northd/0.log" Nov 23 21:20:38 crc kubenswrapper[4726]: I1123 21:20:38.084713 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ff67ad55-4c63-45f3-b6f5-12f941bdd2a7/ovsdbserver-nb/0.log" Nov 23 21:20:38 crc kubenswrapper[4726]: I1123 21:20:38.256417 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ff67ad55-4c63-45f3-b6f5-12f941bdd2a7/openstack-network-exporter/0.log" Nov 23 21:20:38 crc kubenswrapper[4726]: I1123 21:20:38.354218 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1cb13d69-d5a0-4a7b-929b-ed077760725a/openstack-network-exporter/0.log" Nov 23 21:20:38 crc kubenswrapper[4726]: I1123 21:20:38.491406 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1cb13d69-d5a0-4a7b-929b-ed077760725a/ovsdbserver-sb/0.log" Nov 23 21:20:38 crc kubenswrapper[4726]: I1123 21:20:38.772046 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-fbb56fbb6-7fzgk_71d3aaf5-7d1c-46ac-8a80-5ab758b260d4/placement-api/0.log" Nov 23 21:20:38 crc kubenswrapper[4726]: I1123 21:20:38.830294 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-fbb56fbb6-7fzgk_71d3aaf5-7d1c-46ac-8a80-5ab758b260d4/placement-log/0.log" Nov 23 21:20:38 crc kubenswrapper[4726]: I1123 21:20:38.937475 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0883d1c4-a52b-4221-a651-8ba8c5463dd6/setup-container/0.log" Nov 23 21:20:39 crc kubenswrapper[4726]: I1123 21:20:39.310361 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0883d1c4-a52b-4221-a651-8ba8c5463dd6/rabbitmq/0.log" Nov 23 21:20:39 crc kubenswrapper[4726]: I1123 21:20:39.315576 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0883d1c4-a52b-4221-a651-8ba8c5463dd6/setup-container/0.log" Nov 23 21:20:39 crc kubenswrapper[4726]: I1123 21:20:39.335399 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f962af44-2b42-4924-81b8-872c96335906/setup-container/0.log" Nov 23 21:20:39 crc kubenswrapper[4726]: I1123 21:20:39.537112 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f962af44-2b42-4924-81b8-872c96335906/setup-container/0.log" Nov 23 21:20:39 crc kubenswrapper[4726]: I1123 21:20:39.762259 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6_e1af3b69-28d2-4a52-8f13-7cb3e9831fb4/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:39 crc kubenswrapper[4726]: I1123 21:20:39.832932 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f962af44-2b42-4924-81b8-872c96335906/rabbitmq/0.log" Nov 23 21:20:40 crc kubenswrapper[4726]: I1123 21:20:40.105324 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb_dc981c46-b927-4734-a0d1-061826f92146/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:40 crc kubenswrapper[4726]: I1123 21:20:40.362519 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pvpj5_275bc578-25bb-4340-9e5c-f2c296713110/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:40 crc kubenswrapper[4726]: I1123 21:20:40.553837 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mvtwf_80b8ce4c-915f-424b-99c4-a16f0d9a1fc8/ssh-known-hosts-edpm-deployment/0.log" Nov 23 21:20:40 crc kubenswrapper[4726]: I1123 21:20:40.748410 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_efb73145-b382-4a15-89fe-6b74bc9d3594/tempest-tests-tempest-tests-runner/0.log" Nov 23 21:20:41 crc kubenswrapper[4726]: I1123 21:20:41.076150 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_eeaefcf2-c7a8-4b74-b883-9b05a5e01a23/test-operator-logs-container/0.log" Nov 23 21:20:41 crc kubenswrapper[4726]: I1123 21:20:41.358979 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn_da03c2c2-f43a-4f51-9c87-8e2f6b0af367/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:20:41 crc kubenswrapper[4726]: I1123 21:20:41.589533 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:20:41 crc kubenswrapper[4726]: E1123 21:20:41.590787 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:20:52 crc kubenswrapper[4726]: I1123 21:20:52.921267 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_319fe004-3d9c-4bb9-bb8e-114c8949bcc7/memcached/0.log" Nov 23 21:20:56 crc kubenswrapper[4726]: I1123 21:20:56.592742 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:20:56 crc kubenswrapper[4726]: E1123 21:20:56.593506 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:21:08 crc kubenswrapper[4726]: I1123 21:21:08.592419 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:21:08 crc kubenswrapper[4726]: E1123 21:21:08.593502 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:21:11 crc kubenswrapper[4726]: I1123 21:21:11.537510 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/util/0.log" Nov 23 21:21:11 crc kubenswrapper[4726]: I1123 21:21:11.741505 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/util/0.log" Nov 23 21:21:11 crc kubenswrapper[4726]: I1123 21:21:11.821131 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/pull/0.log" Nov 23 21:21:11 crc kubenswrapper[4726]: I1123 21:21:11.825402 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/pull/0.log" Nov 23 21:21:11 crc kubenswrapper[4726]: I1123 21:21:11.994691 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/util/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.001048 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/extract/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.035259 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/pull/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.186202 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-q7qxp_d481bfb6-f28c-49ed-8efe-7fc35dc6d608/kube-rbac-proxy/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.260008 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-q7qxp_d481bfb6-f28c-49ed-8efe-7fc35dc6d608/manager/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.292178 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-xt9t6_a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7/kube-rbac-proxy/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.480810 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-xt9t6_a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7/manager/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.534813 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-wh58z_2e4ac3d0-65ba-4140-9a03-c269dfebad67/kube-rbac-proxy/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.588647 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-wh58z_2e4ac3d0-65ba-4140-9a03-c269dfebad67/manager/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.756377 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-6d8lj_bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb/kube-rbac-proxy/0.log" Nov 23 21:21:12 crc kubenswrapper[4726]: I1123 21:21:12.785262 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-6d8lj_bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb/manager/0.log" Nov 23 21:21:13 crc kubenswrapper[4726]: I1123 21:21:13.185854 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-jc55l_9c05c149-b530-4ca3-b201-309826c5e095/kube-rbac-proxy/0.log" Nov 23 21:21:13 crc kubenswrapper[4726]: I1123 21:21:13.276819 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-jc55l_9c05c149-b530-4ca3-b201-309826c5e095/manager/0.log" Nov 23 21:21:13 crc kubenswrapper[4726]: I1123 21:21:13.361400 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-574zb_c3279a0f-e448-47b2-91b3-6de329d77cec/kube-rbac-proxy/0.log" Nov 23 21:21:13 crc kubenswrapper[4726]: I1123 21:21:13.406423 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-574zb_c3279a0f-e448-47b2-91b3-6de329d77cec/manager/0.log" Nov 23 21:21:13 crc kubenswrapper[4726]: I1123 21:21:13.547794 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-w4lxk_765e2946-d293-4840-baf2-96a478748a25/kube-rbac-proxy/0.log" Nov 23 21:21:13 crc kubenswrapper[4726]: I1123 21:21:13.730373 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-w4lxk_765e2946-d293-4840-baf2-96a478748a25/manager/0.log" Nov 23 21:21:13 crc kubenswrapper[4726]: I1123 21:21:13.760123 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-m4xc4_c53c8659-c283-4c78-88b4-b60f8b040dd7/kube-rbac-proxy/0.log" Nov 23 21:21:13 crc kubenswrapper[4726]: I1123 21:21:13.872748 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-m4xc4_c53c8659-c283-4c78-88b4-b60f8b040dd7/manager/0.log" Nov 23 21:21:14 crc kubenswrapper[4726]: I1123 21:21:14.050926 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-s7746_76e7ee34-d8e7-4d5b-b79e-acd11b2b0735/kube-rbac-proxy/0.log" Nov 23 21:21:14 crc kubenswrapper[4726]: I1123 21:21:14.121380 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-5s9v2_a653d1f5-fe23-4a55-9422-33964eeb8488/kube-rbac-proxy/0.log" Nov 23 21:21:14 crc kubenswrapper[4726]: I1123 21:21:14.134642 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-s7746_76e7ee34-d8e7-4d5b-b79e-acd11b2b0735/manager/0.log" Nov 23 21:21:14 crc kubenswrapper[4726]: I1123 21:21:14.738023 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-8dz5d_b89cb12d-b390-40ed-908f-e1095fe0c62b/kube-rbac-proxy/0.log" Nov 23 21:21:14 crc kubenswrapper[4726]: I1123 21:21:14.775176 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-8dz5d_b89cb12d-b390-40ed-908f-e1095fe0c62b/manager/0.log" Nov 23 21:21:14 crc kubenswrapper[4726]: I1123 21:21:14.831646 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-5s9v2_a653d1f5-fe23-4a55-9422-33964eeb8488/manager/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.001980 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-4p7kg_4301d05b-4769-4c53-b557-bd6aa37efa68/kube-rbac-proxy/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.042655 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-4p7kg_4301d05b-4769-4c53-b557-bd6aa37efa68/manager/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.152521 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-qvx6m_384a1b22-97f2-4f0b-ad98-52adbbf5178b/kube-rbac-proxy/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.312551 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-qvx6m_384a1b22-97f2-4f0b-ad98-52adbbf5178b/manager/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.383354 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-7lwwg_20b0ca2b-a3ee-45c9-8b34-b476245a4ed6/manager/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.418911 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-7lwwg_20b0ca2b-a3ee-45c9-8b34-b476245a4ed6/kube-rbac-proxy/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.515818 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44hnx49_ce7ed28d-d56b-44de-b207-7c47ab684921/kube-rbac-proxy/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.593906 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44hnx49_ce7ed28d-d56b-44de-b207-7c47ab684921/manager/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.743991 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-t9ptk_706d33da-dac2-4403-a4d8-d8e7f04f772b/kube-rbac-proxy/0.log" Nov 23 21:21:15 crc kubenswrapper[4726]: I1123 21:21:15.968481 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-v5bmf_2a88699d-09b3-458d-bbbd-03834f3c6978/kube-rbac-proxy/0.log" Nov 23 21:21:16 crc kubenswrapper[4726]: I1123 21:21:16.168525 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-v5bmf_2a88699d-09b3-458d-bbbd-03834f3c6978/operator/0.log" Nov 23 21:21:16 crc kubenswrapper[4726]: I1123 21:21:16.287046 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-whxfc_224802b7-2607-4e77-95e0-d7e02023c20b/registry-server/0.log" Nov 23 21:21:16 crc kubenswrapper[4726]: I1123 21:21:16.463033 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-2k5zx_6e692743-eb35-4a91-97af-8d86cf694470/kube-rbac-proxy/0.log" Nov 23 21:21:16 crc kubenswrapper[4726]: I1123 21:21:16.723111 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-p7r99_42d03b10-01f7-4963-a97b-f3d443a1cdf7/kube-rbac-proxy/0.log" Nov 23 21:21:16 crc kubenswrapper[4726]: I1123 21:21:16.731856 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-p7r99_42d03b10-01f7-4963-a97b-f3d443a1cdf7/manager/0.log" Nov 23 21:21:16 crc kubenswrapper[4726]: I1123 21:21:16.748050 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-2k5zx_6e692743-eb35-4a91-97af-8d86cf694470/manager/0.log" Nov 23 21:21:16 crc kubenswrapper[4726]: I1123 21:21:16.834410 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-t9ptk_706d33da-dac2-4403-a4d8-d8e7f04f772b/manager/0.log" Nov 23 21:21:16 crc kubenswrapper[4726]: I1123 21:21:16.992829 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-frss6_cfdab430-7dd4-4612-a65b-b381b5b58115/kube-rbac-proxy/0.log" Nov 23 21:21:17 crc kubenswrapper[4726]: I1123 21:21:17.003712 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv_e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9/operator/0.log" Nov 23 21:21:17 crc kubenswrapper[4726]: I1123 21:21:17.064671 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-frss6_cfdab430-7dd4-4612-a65b-b381b5b58115/manager/0.log" Nov 23 21:21:17 crc kubenswrapper[4726]: I1123 21:21:17.159353 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-527bk_74d488f0-9a23-4774-94e7-75605573ce43/kube-rbac-proxy/0.log" Nov 23 21:21:17 crc kubenswrapper[4726]: I1123 21:21:17.259370 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-527bk_74d488f0-9a23-4774-94e7-75605573ce43/manager/0.log" Nov 23 21:21:17 crc kubenswrapper[4726]: I1123 21:21:17.288381 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-5w2nt_c22e9617-abf8-4370-ba10-31e967257270/kube-rbac-proxy/0.log" Nov 23 21:21:17 crc kubenswrapper[4726]: I1123 21:21:17.395944 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-5w2nt_c22e9617-abf8-4370-ba10-31e967257270/manager/0.log" Nov 23 21:21:17 crc kubenswrapper[4726]: I1123 21:21:17.504523 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-7x7b7_023e8819-c436-4592-a6f6-ea3448733363/manager/0.log" Nov 23 21:21:17 crc kubenswrapper[4726]: I1123 21:21:17.506157 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-7x7b7_023e8819-c436-4592-a6f6-ea3448733363/kube-rbac-proxy/0.log" Nov 23 21:21:19 crc kubenswrapper[4726]: I1123 21:21:19.589089 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:21:19 crc kubenswrapper[4726]: E1123 21:21:19.589646 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:21:21 crc kubenswrapper[4726]: I1123 21:21:21.528532 4726 scope.go:117] "RemoveContainer" containerID="68f10f5fc1b79051213dc60029b4fcaf18d2a161474dee448f0f6a8ddf828d74" Nov 23 21:21:21 crc kubenswrapper[4726]: I1123 21:21:21.549280 4726 scope.go:117] "RemoveContainer" containerID="51a232cff9cd0e2817c67447b8205fbc735bc31d54ab965d13526ddf7519db2e" Nov 23 21:21:21 crc kubenswrapper[4726]: I1123 21:21:21.573179 4726 scope.go:117] "RemoveContainer" containerID="9c3e3fd52204d39ef481b804fc5cac3a9d03b6a328dec9f14ed20797b5ade051" Nov 23 21:21:31 crc kubenswrapper[4726]: I1123 21:21:31.589024 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:21:31 crc kubenswrapper[4726]: E1123 21:21:31.589954 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:21:35 crc kubenswrapper[4726]: I1123 21:21:35.533082 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-hlvv6_4ff399e8-7c2b-497c-aec8-96383b2f92fb/control-plane-machine-set-operator/0.log" Nov 23 21:21:35 crc kubenswrapper[4726]: I1123 21:21:35.726009 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vfgwr_002cee03-8433-4e0d-b648-45300fadd459/kube-rbac-proxy/0.log" Nov 23 21:21:35 crc kubenswrapper[4726]: I1123 21:21:35.751356 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vfgwr_002cee03-8433-4e0d-b648-45300fadd459/machine-api-operator/0.log" Nov 23 21:21:42 crc kubenswrapper[4726]: I1123 21:21:42.596142 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:21:42 crc kubenswrapper[4726]: E1123 21:21:42.598322 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:21:48 crc kubenswrapper[4726]: I1123 21:21:48.977352 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vp4hq_3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd/cert-manager-controller/0.log" Nov 23 21:21:49 crc kubenswrapper[4726]: I1123 21:21:49.194980 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-llcjm_d7287208-3876-4ffe-b893-a588d7383fe8/cert-manager-webhook/0.log" Nov 23 21:21:49 crc kubenswrapper[4726]: I1123 21:21:49.229235 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-dcnpn_b67043f1-eb7f-4239-b3a2-1f328e5f681f/cert-manager-cainjector/0.log" Nov 23 21:21:55 crc kubenswrapper[4726]: I1123 21:21:55.589856 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:21:55 crc kubenswrapper[4726]: E1123 21:21:55.590670 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:22:04 crc kubenswrapper[4726]: I1123 21:22:04.249305 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-98ljz_6dc6c946-84ef-4ba1-8753-efd769c877c0/nmstate-console-plugin/0.log" Nov 23 21:22:04 crc kubenswrapper[4726]: I1123 21:22:04.453987 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-w9sld_97c85373-c730-40fd-8b73-0b163c59e6ba/nmstate-handler/0.log" Nov 23 21:22:04 crc kubenswrapper[4726]: I1123 21:22:04.549846 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-pnnn6_fdc047b2-c091-49cc-8a3c-9663b20e7cd1/kube-rbac-proxy/0.log" Nov 23 21:22:04 crc kubenswrapper[4726]: I1123 21:22:04.585268 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-pnnn6_fdc047b2-c091-49cc-8a3c-9663b20e7cd1/nmstate-metrics/0.log" Nov 23 21:22:04 crc kubenswrapper[4726]: I1123 21:22:04.738348 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-w8pws_0debdf86-4fa6-46aa-8ffc-513dbb889153/nmstate-operator/0.log" Nov 23 21:22:04 crc kubenswrapper[4726]: I1123 21:22:04.851691 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-jflhl_f89596fa-328e-483b-a5dd-34a241972f63/nmstate-webhook/0.log" Nov 23 21:22:08 crc kubenswrapper[4726]: I1123 21:22:08.590283 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:22:08 crc kubenswrapper[4726]: E1123 21:22:08.591092 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:22:22 crc kubenswrapper[4726]: I1123 21:22:22.839089 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-2vb27_dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0/controller/0.log" Nov 23 21:22:22 crc kubenswrapper[4726]: I1123 21:22:22.874341 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-2vb27_dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0/kube-rbac-proxy/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.052237 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-frr-files/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.214577 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-metrics/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.241766 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-frr-files/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.291071 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-reloader/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.324390 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-reloader/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.452124 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-frr-files/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.515964 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-reloader/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.530811 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-metrics/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.539504 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-metrics/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.589322 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:22:23 crc kubenswrapper[4726]: E1123 21:22:23.589579 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.778582 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-reloader/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.810263 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-metrics/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.851170 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-frr-files/0.log" Nov 23 21:22:23 crc kubenswrapper[4726]: I1123 21:22:23.853699 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/controller/0.log" Nov 23 21:22:24 crc kubenswrapper[4726]: I1123 21:22:24.070034 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/frr-metrics/0.log" Nov 23 21:22:24 crc kubenswrapper[4726]: I1123 21:22:24.140690 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/kube-rbac-proxy/0.log" Nov 23 21:22:24 crc kubenswrapper[4726]: I1123 21:22:24.172951 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/kube-rbac-proxy-frr/0.log" Nov 23 21:22:24 crc kubenswrapper[4726]: I1123 21:22:24.344461 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/reloader/0.log" Nov 23 21:22:24 crc kubenswrapper[4726]: I1123 21:22:24.502581 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-mlnqp_3130c09b-7002-4448-bdd4-87b73f70ff99/frr-k8s-webhook-server/0.log" Nov 23 21:22:24 crc kubenswrapper[4726]: I1123 21:22:24.734503 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-64f7f896bc-tb44t_be91600b-06c3-4acc-a0d6-d7e732a00228/manager/0.log" Nov 23 21:22:24 crc kubenswrapper[4726]: I1123 21:22:24.986967 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c66969f79-8m676_bf3b7cb9-1dc8-4966-b15c-71d77aa2751b/webhook-server/0.log" Nov 23 21:22:25 crc kubenswrapper[4726]: I1123 21:22:25.079929 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zkvn5_d99ab20e-6022-4d7e-8ff2-571cd0961a08/kube-rbac-proxy/0.log" Nov 23 21:22:25 crc kubenswrapper[4726]: I1123 21:22:25.425149 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/frr/0.log" Nov 23 21:22:25 crc kubenswrapper[4726]: I1123 21:22:25.615564 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zkvn5_d99ab20e-6022-4d7e-8ff2-571cd0961a08/speaker/0.log" Nov 23 21:22:38 crc kubenswrapper[4726]: I1123 21:22:38.588824 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:22:38 crc kubenswrapper[4726]: E1123 21:22:38.589516 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:22:40 crc kubenswrapper[4726]: I1123 21:22:40.812732 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/util/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.128829 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/util/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.156397 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/pull/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.167555 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/pull/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.267877 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/util/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.400638 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/pull/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.412694 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/extract/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.518141 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-utilities/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.744698 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-content/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.809380 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-utilities/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.853269 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-content/0.log" Nov 23 21:22:41 crc kubenswrapper[4726]: I1123 21:22:41.955301 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-utilities/0.log" Nov 23 21:22:42 crc kubenswrapper[4726]: I1123 21:22:42.101108 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-content/0.log" Nov 23 21:22:42 crc kubenswrapper[4726]: I1123 21:22:42.197677 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-utilities/0.log" Nov 23 21:22:42 crc kubenswrapper[4726]: I1123 21:22:42.605316 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/registry-server/0.log" Nov 23 21:22:42 crc kubenswrapper[4726]: I1123 21:22:42.615610 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-utilities/0.log" Nov 23 21:22:42 crc kubenswrapper[4726]: I1123 21:22:42.633976 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-content/0.log" Nov 23 21:22:42 crc kubenswrapper[4726]: I1123 21:22:42.654293 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-content/0.log" Nov 23 21:22:42 crc kubenswrapper[4726]: I1123 21:22:42.816073 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-utilities/0.log" Nov 23 21:22:42 crc kubenswrapper[4726]: I1123 21:22:42.855469 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-content/0.log" Nov 23 21:22:43 crc kubenswrapper[4726]: I1123 21:22:43.137881 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/util/0.log" Nov 23 21:22:43 crc kubenswrapper[4726]: I1123 21:22:43.419175 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/pull/0.log" Nov 23 21:22:43 crc kubenswrapper[4726]: I1123 21:22:43.470028 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/registry-server/0.log" Nov 23 21:22:43 crc kubenswrapper[4726]: I1123 21:22:43.518373 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/util/0.log" Nov 23 21:22:43 crc kubenswrapper[4726]: I1123 21:22:43.555023 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/pull/0.log" Nov 23 21:22:43 crc kubenswrapper[4726]: I1123 21:22:43.750223 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/pull/0.log" Nov 23 21:22:43 crc kubenswrapper[4726]: I1123 21:22:43.776044 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/util/0.log" Nov 23 21:22:43 crc kubenswrapper[4726]: I1123 21:22:43.892634 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/extract/0.log" Nov 23 21:22:44 crc kubenswrapper[4726]: I1123 21:22:44.019541 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7rvp2_cfd53c53-6459-4a5d-b8e4-bf47ebb6da34/marketplace-operator/0.log" Nov 23 21:22:44 crc kubenswrapper[4726]: I1123 21:22:44.064390 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-utilities/0.log" Nov 23 21:22:44 crc kubenswrapper[4726]: I1123 21:22:44.286179 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-content/0.log" Nov 23 21:22:44 crc kubenswrapper[4726]: I1123 21:22:44.322387 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-utilities/0.log" Nov 23 21:22:44 crc kubenswrapper[4726]: I1123 21:22:44.364727 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-content/0.log" Nov 23 21:22:44 crc kubenswrapper[4726]: I1123 21:22:44.951834 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-content/0.log" Nov 23 21:22:45 crc kubenswrapper[4726]: I1123 21:22:45.031635 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-utilities/0.log" Nov 23 21:22:45 crc kubenswrapper[4726]: I1123 21:22:45.046000 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/registry-server/0.log" Nov 23 21:22:45 crc kubenswrapper[4726]: I1123 21:22:45.178083 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-utilities/0.log" Nov 23 21:22:45 crc kubenswrapper[4726]: I1123 21:22:45.438542 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-content/0.log" Nov 23 21:22:45 crc kubenswrapper[4726]: I1123 21:22:45.439483 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-content/0.log" Nov 23 21:22:45 crc kubenswrapper[4726]: I1123 21:22:45.443384 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-utilities/0.log" Nov 23 21:22:46 crc kubenswrapper[4726]: I1123 21:22:46.167775 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-utilities/0.log" Nov 23 21:22:46 crc kubenswrapper[4726]: I1123 21:22:46.278838 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-content/0.log" Nov 23 21:22:46 crc kubenswrapper[4726]: I1123 21:22:46.795625 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/registry-server/0.log" Nov 23 21:22:50 crc kubenswrapper[4726]: I1123 21:22:50.589176 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:22:50 crc kubenswrapper[4726]: E1123 21:22:50.590126 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:23:05 crc kubenswrapper[4726]: I1123 21:23:05.589111 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:23:05 crc kubenswrapper[4726]: E1123 21:23:05.590059 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:23:10 crc kubenswrapper[4726]: I1123 21:23:10.967280 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dt5qh"] Nov 23 21:23:10 crc kubenswrapper[4726]: E1123 21:23:10.968185 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c594fef9-43cd-4693-9f79-6ab4d8a8b5fd" containerName="container-00" Nov 23 21:23:10 crc kubenswrapper[4726]: I1123 21:23:10.968199 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c594fef9-43cd-4693-9f79-6ab4d8a8b5fd" containerName="container-00" Nov 23 21:23:10 crc kubenswrapper[4726]: I1123 21:23:10.968422 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c594fef9-43cd-4693-9f79-6ab4d8a8b5fd" containerName="container-00" Nov 23 21:23:10 crc kubenswrapper[4726]: I1123 21:23:10.969763 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.048511 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dt5qh"] Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.080437 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt8p9\" (UniqueName: \"kubernetes.io/projected/15980268-ae79-4bc6-9c79-7cf8c1f39e37-kube-api-access-dt8p9\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.080581 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-utilities\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.080625 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-catalog-content\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.181767 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-utilities\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.181834 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-catalog-content\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.181917 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt8p9\" (UniqueName: \"kubernetes.io/projected/15980268-ae79-4bc6-9c79-7cf8c1f39e37-kube-api-access-dt8p9\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.182478 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-utilities\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.182519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-catalog-content\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.230703 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt8p9\" (UniqueName: \"kubernetes.io/projected/15980268-ae79-4bc6-9c79-7cf8c1f39e37-kube-api-access-dt8p9\") pod \"certified-operators-dt5qh\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:11 crc kubenswrapper[4726]: I1123 21:23:11.289343 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:12 crc kubenswrapper[4726]: I1123 21:23:12.428676 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dt5qh"] Nov 23 21:23:12 crc kubenswrapper[4726]: I1123 21:23:12.533568 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt5qh" event={"ID":"15980268-ae79-4bc6-9c79-7cf8c1f39e37","Type":"ContainerStarted","Data":"40872ebcc7acd0592b55d3d358907bf17df0a6f44e8bcfe72c30e268d5bd5174"} Nov 23 21:23:13 crc kubenswrapper[4726]: I1123 21:23:13.542306 4726 generic.go:334] "Generic (PLEG): container finished" podID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerID="c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17" exitCode=0 Nov 23 21:23:13 crc kubenswrapper[4726]: I1123 21:23:13.542345 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt5qh" event={"ID":"15980268-ae79-4bc6-9c79-7cf8c1f39e37","Type":"ContainerDied","Data":"c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17"} Nov 23 21:23:13 crc kubenswrapper[4726]: I1123 21:23:13.545230 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 21:23:14 crc kubenswrapper[4726]: I1123 21:23:14.552011 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt5qh" event={"ID":"15980268-ae79-4bc6-9c79-7cf8c1f39e37","Type":"ContainerStarted","Data":"4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39"} Nov 23 21:23:16 crc kubenswrapper[4726]: I1123 21:23:16.571939 4726 generic.go:334] "Generic (PLEG): container finished" podID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerID="4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39" exitCode=0 Nov 23 21:23:16 crc kubenswrapper[4726]: I1123 21:23:16.572022 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt5qh" event={"ID":"15980268-ae79-4bc6-9c79-7cf8c1f39e37","Type":"ContainerDied","Data":"4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39"} Nov 23 21:23:17 crc kubenswrapper[4726]: I1123 21:23:17.581768 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt5qh" event={"ID":"15980268-ae79-4bc6-9c79-7cf8c1f39e37","Type":"ContainerStarted","Data":"13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66"} Nov 23 21:23:17 crc kubenswrapper[4726]: I1123 21:23:17.600646 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dt5qh" podStartSLOduration=4.167270933 podStartE2EDuration="7.600630044s" podCreationTimestamp="2025-11-23 21:23:10 +0000 UTC" firstStartedPulling="2025-11-23 21:23:13.545013631 +0000 UTC m=+4501.694054587" lastFinishedPulling="2025-11-23 21:23:16.978372742 +0000 UTC m=+4505.127413698" observedRunningTime="2025-11-23 21:23:17.596890639 +0000 UTC m=+4505.745931605" watchObservedRunningTime="2025-11-23 21:23:17.600630044 +0000 UTC m=+4505.749671000" Nov 23 21:23:18 crc kubenswrapper[4726]: I1123 21:23:18.589049 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:23:18 crc kubenswrapper[4726]: E1123 21:23:18.589331 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:23:21 crc kubenswrapper[4726]: I1123 21:23:21.289522 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:21 crc kubenswrapper[4726]: I1123 21:23:21.289990 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:21 crc kubenswrapper[4726]: I1123 21:23:21.351106 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:30 crc kubenswrapper[4726]: I1123 21:23:30.589716 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:23:30 crc kubenswrapper[4726]: E1123 21:23:30.590749 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:23:31 crc kubenswrapper[4726]: I1123 21:23:31.334581 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:31 crc kubenswrapper[4726]: I1123 21:23:31.406519 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dt5qh"] Nov 23 21:23:31 crc kubenswrapper[4726]: I1123 21:23:31.701222 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dt5qh" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerName="registry-server" containerID="cri-o://13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66" gracePeriod=2 Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.225235 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.301246 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt8p9\" (UniqueName: \"kubernetes.io/projected/15980268-ae79-4bc6-9c79-7cf8c1f39e37-kube-api-access-dt8p9\") pod \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.302079 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-catalog-content\") pod \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.302287 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-utilities\") pod \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\" (UID: \"15980268-ae79-4bc6-9c79-7cf8c1f39e37\") " Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.302887 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-utilities" (OuterVolumeSpecName: "utilities") pod "15980268-ae79-4bc6-9c79-7cf8c1f39e37" (UID: "15980268-ae79-4bc6-9c79-7cf8c1f39e37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.304587 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.324680 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15980268-ae79-4bc6-9c79-7cf8c1f39e37-kube-api-access-dt8p9" (OuterVolumeSpecName: "kube-api-access-dt8p9") pod "15980268-ae79-4bc6-9c79-7cf8c1f39e37" (UID: "15980268-ae79-4bc6-9c79-7cf8c1f39e37"). InnerVolumeSpecName "kube-api-access-dt8p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.376964 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15980268-ae79-4bc6-9c79-7cf8c1f39e37" (UID: "15980268-ae79-4bc6-9c79-7cf8c1f39e37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.406011 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt8p9\" (UniqueName: \"kubernetes.io/projected/15980268-ae79-4bc6-9c79-7cf8c1f39e37-kube-api-access-dt8p9\") on node \"crc\" DevicePath \"\"" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.406040 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15980268-ae79-4bc6-9c79-7cf8c1f39e37-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.727442 4726 generic.go:334] "Generic (PLEG): container finished" podID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerID="13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66" exitCode=0 Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.727492 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt5qh" event={"ID":"15980268-ae79-4bc6-9c79-7cf8c1f39e37","Type":"ContainerDied","Data":"13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66"} Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.727524 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dt5qh" event={"ID":"15980268-ae79-4bc6-9c79-7cf8c1f39e37","Type":"ContainerDied","Data":"40872ebcc7acd0592b55d3d358907bf17df0a6f44e8bcfe72c30e268d5bd5174"} Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.727541 4726 scope.go:117] "RemoveContainer" containerID="13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.727544 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dt5qh" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.766451 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dt5qh"] Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.775126 4726 scope.go:117] "RemoveContainer" containerID="4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.784258 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dt5qh"] Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.803126 4726 scope.go:117] "RemoveContainer" containerID="c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.871572 4726 scope.go:117] "RemoveContainer" containerID="13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66" Nov 23 21:23:32 crc kubenswrapper[4726]: E1123 21:23:32.872459 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66\": container with ID starting with 13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66 not found: ID does not exist" containerID="13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.872584 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66"} err="failed to get container status \"13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66\": rpc error: code = NotFound desc = could not find container \"13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66\": container with ID starting with 13bc66a6aab2a4a33893522d5e8d2ced73346271790d02cac85f4c6327b88a66 not found: ID does not exist" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.872702 4726 scope.go:117] "RemoveContainer" containerID="4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39" Nov 23 21:23:32 crc kubenswrapper[4726]: E1123 21:23:32.873391 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39\": container with ID starting with 4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39 not found: ID does not exist" containerID="4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.873555 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39"} err="failed to get container status \"4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39\": rpc error: code = NotFound desc = could not find container \"4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39\": container with ID starting with 4317bc39856efa0b527e25ead6230644dc45fe8d8012f341237f5b96a767cd39 not found: ID does not exist" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.873599 4726 scope.go:117] "RemoveContainer" containerID="c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17" Nov 23 21:23:32 crc kubenswrapper[4726]: E1123 21:23:32.874184 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17\": container with ID starting with c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17 not found: ID does not exist" containerID="c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17" Nov 23 21:23:32 crc kubenswrapper[4726]: I1123 21:23:32.874238 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17"} err="failed to get container status \"c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17\": rpc error: code = NotFound desc = could not find container \"c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17\": container with ID starting with c9274efcea1038d2098219bca674901c78b810f4c7a3da893318d1a66f4a7d17 not found: ID does not exist" Nov 23 21:23:34 crc kubenswrapper[4726]: I1123 21:23:34.603603 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" path="/var/lib/kubelet/pods/15980268-ae79-4bc6-9c79-7cf8c1f39e37/volumes" Nov 23 21:23:44 crc kubenswrapper[4726]: I1123 21:23:44.589307 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:23:44 crc kubenswrapper[4726]: E1123 21:23:44.593342 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:23:57 crc kubenswrapper[4726]: I1123 21:23:57.590103 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:23:57 crc kubenswrapper[4726]: E1123 21:23:57.590960 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:24:08 crc kubenswrapper[4726]: I1123 21:24:08.588923 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:24:08 crc kubenswrapper[4726]: E1123 21:24:08.589650 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:24:20 crc kubenswrapper[4726]: I1123 21:24:20.590295 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:24:21 crc kubenswrapper[4726]: I1123 21:24:21.226025 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"d78047930e4aba8b59f1ecbdc499155597ea2347de6f9220fe2147928a3ef4b3"} Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.388516 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j9vwx"] Nov 23 21:24:53 crc kubenswrapper[4726]: E1123 21:24:53.389477 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerName="registry-server" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.389492 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerName="registry-server" Nov 23 21:24:53 crc kubenswrapper[4726]: E1123 21:24:53.389521 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerName="extract-content" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.389529 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerName="extract-content" Nov 23 21:24:53 crc kubenswrapper[4726]: E1123 21:24:53.389547 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerName="extract-utilities" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.389557 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerName="extract-utilities" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.389769 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="15980268-ae79-4bc6-9c79-7cf8c1f39e37" containerName="registry-server" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.391347 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.414411 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9vwx"] Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.524811 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxfgz\" (UniqueName: \"kubernetes.io/projected/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-kube-api-access-cxfgz\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.525058 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-utilities\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.525124 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-catalog-content\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.627371 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-utilities\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.627417 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-catalog-content\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.627558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxfgz\" (UniqueName: \"kubernetes.io/projected/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-kube-api-access-cxfgz\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.627840 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-utilities\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.627961 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-catalog-content\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.661135 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxfgz\" (UniqueName: \"kubernetes.io/projected/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-kube-api-access-cxfgz\") pod \"redhat-operators-j9vwx\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:53 crc kubenswrapper[4726]: I1123 21:24:53.716118 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:24:54 crc kubenswrapper[4726]: I1123 21:24:54.262745 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9vwx"] Nov 23 21:24:54 crc kubenswrapper[4726]: I1123 21:24:54.560296 4726 generic.go:334] "Generic (PLEG): container finished" podID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerID="a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9" exitCode=0 Nov 23 21:24:54 crc kubenswrapper[4726]: I1123 21:24:54.560335 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9vwx" event={"ID":"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef","Type":"ContainerDied","Data":"a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9"} Nov 23 21:24:54 crc kubenswrapper[4726]: I1123 21:24:54.560359 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9vwx" event={"ID":"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef","Type":"ContainerStarted","Data":"31678dbb77fc30cfc6776448939c3de51d8fad4eea51dcdaa10f746531b8a4a8"} Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.571466 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9vwx" event={"ID":"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef","Type":"ContainerStarted","Data":"0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82"} Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.784417 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w4bmb"] Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.793218 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.825717 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w4bmb"] Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.874714 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-catalog-content\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.874812 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-utilities\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.874939 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t2gq\" (UniqueName: \"kubernetes.io/projected/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-kube-api-access-9t2gq\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.976537 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t2gq\" (UniqueName: \"kubernetes.io/projected/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-kube-api-access-9t2gq\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.976698 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-catalog-content\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.976744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-utilities\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.977224 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-utilities\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.977454 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-catalog-content\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:55 crc kubenswrapper[4726]: I1123 21:24:55.995832 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t2gq\" (UniqueName: \"kubernetes.io/projected/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-kube-api-access-9t2gq\") pod \"community-operators-w4bmb\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:56 crc kubenswrapper[4726]: I1123 21:24:56.124470 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:24:56 crc kubenswrapper[4726]: I1123 21:24:56.618509 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w4bmb"] Nov 23 21:24:57 crc kubenswrapper[4726]: I1123 21:24:57.597686 4726 generic.go:334] "Generic (PLEG): container finished" podID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerID="34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2" exitCode=0 Nov 23 21:24:57 crc kubenswrapper[4726]: I1123 21:24:57.597736 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4bmb" event={"ID":"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961","Type":"ContainerDied","Data":"34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2"} Nov 23 21:24:57 crc kubenswrapper[4726]: I1123 21:24:57.597767 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4bmb" event={"ID":"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961","Type":"ContainerStarted","Data":"8d126dbf0e3e3df37c9bdf9b506c82eb3b1fc2aeeeaaa756c8d8e45c806974dc"} Nov 23 21:24:58 crc kubenswrapper[4726]: I1123 21:24:58.605807 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4bmb" event={"ID":"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961","Type":"ContainerStarted","Data":"217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc"} Nov 23 21:24:59 crc kubenswrapper[4726]: I1123 21:24:59.620459 4726 generic.go:334] "Generic (PLEG): container finished" podID="d5709a4c-d883-4f50-b622-f482770107ff" containerID="4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce" exitCode=0 Nov 23 21:24:59 crc kubenswrapper[4726]: I1123 21:24:59.620530 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" event={"ID":"d5709a4c-d883-4f50-b622-f482770107ff","Type":"ContainerDied","Data":"4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce"} Nov 23 21:24:59 crc kubenswrapper[4726]: I1123 21:24:59.622166 4726 scope.go:117] "RemoveContainer" containerID="4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce" Nov 23 21:24:59 crc kubenswrapper[4726]: I1123 21:24:59.625729 4726 generic.go:334] "Generic (PLEG): container finished" podID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerID="0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82" exitCode=0 Nov 23 21:24:59 crc kubenswrapper[4726]: I1123 21:24:59.625828 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9vwx" event={"ID":"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef","Type":"ContainerDied","Data":"0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82"} Nov 23 21:25:00 crc kubenswrapper[4726]: I1123 21:25:00.589454 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gz4k6_must-gather-cw2lq_d5709a4c-d883-4f50-b622-f482770107ff/gather/0.log" Nov 23 21:25:00 crc kubenswrapper[4726]: I1123 21:25:00.639335 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9vwx" event={"ID":"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef","Type":"ContainerStarted","Data":"35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2"} Nov 23 21:25:00 crc kubenswrapper[4726]: I1123 21:25:00.642583 4726 generic.go:334] "Generic (PLEG): container finished" podID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerID="217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc" exitCode=0 Nov 23 21:25:00 crc kubenswrapper[4726]: I1123 21:25:00.642655 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4bmb" event={"ID":"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961","Type":"ContainerDied","Data":"217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc"} Nov 23 21:25:00 crc kubenswrapper[4726]: I1123 21:25:00.674901 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j9vwx" podStartSLOduration=2.070983546 podStartE2EDuration="7.67488404s" podCreationTimestamp="2025-11-23 21:24:53 +0000 UTC" firstStartedPulling="2025-11-23 21:24:54.561933544 +0000 UTC m=+4602.710974500" lastFinishedPulling="2025-11-23 21:25:00.165834048 +0000 UTC m=+4608.314874994" observedRunningTime="2025-11-23 21:25:00.664172869 +0000 UTC m=+4608.813213825" watchObservedRunningTime="2025-11-23 21:25:00.67488404 +0000 UTC m=+4608.823924996" Nov 23 21:25:01 crc kubenswrapper[4726]: I1123 21:25:01.653562 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4bmb" event={"ID":"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961","Type":"ContainerStarted","Data":"0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f"} Nov 23 21:25:01 crc kubenswrapper[4726]: I1123 21:25:01.685993 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w4bmb" podStartSLOduration=3.202210378 podStartE2EDuration="6.685976836s" podCreationTimestamp="2025-11-23 21:24:55 +0000 UTC" firstStartedPulling="2025-11-23 21:24:57.60143466 +0000 UTC m=+4605.750475656" lastFinishedPulling="2025-11-23 21:25:01.085201168 +0000 UTC m=+4609.234242114" observedRunningTime="2025-11-23 21:25:01.681564762 +0000 UTC m=+4609.830605728" watchObservedRunningTime="2025-11-23 21:25:01.685976836 +0000 UTC m=+4609.835017792" Nov 23 21:25:03 crc kubenswrapper[4726]: I1123 21:25:03.716755 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:25:03 crc kubenswrapper[4726]: I1123 21:25:03.717111 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:25:04 crc kubenswrapper[4726]: I1123 21:25:04.841225 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j9vwx" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="registry-server" probeResult="failure" output=< Nov 23 21:25:04 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 21:25:04 crc kubenswrapper[4726]: > Nov 23 21:25:06 crc kubenswrapper[4726]: I1123 21:25:06.125121 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:25:06 crc kubenswrapper[4726]: I1123 21:25:06.126624 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:25:06 crc kubenswrapper[4726]: I1123 21:25:06.202180 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:25:06 crc kubenswrapper[4726]: I1123 21:25:06.784441 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:25:06 crc kubenswrapper[4726]: I1123 21:25:06.845704 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w4bmb"] Nov 23 21:25:08 crc kubenswrapper[4726]: I1123 21:25:08.748400 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w4bmb" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerName="registry-server" containerID="cri-o://0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f" gracePeriod=2 Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.281891 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.394493 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-utilities\") pod \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.394655 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t2gq\" (UniqueName: \"kubernetes.io/projected/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-kube-api-access-9t2gq\") pod \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.394725 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-catalog-content\") pod \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\" (UID: \"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961\") " Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.395507 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-utilities" (OuterVolumeSpecName: "utilities") pod "d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" (UID: "d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.405081 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-kube-api-access-9t2gq" (OuterVolumeSpecName: "kube-api-access-9t2gq") pod "d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" (UID: "d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961"). InnerVolumeSpecName "kube-api-access-9t2gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.442253 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" (UID: "d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.498168 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.498456 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t2gq\" (UniqueName: \"kubernetes.io/projected/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-kube-api-access-9t2gq\") on node \"crc\" DevicePath \"\"" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.498525 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.758633 4726 generic.go:334] "Generic (PLEG): container finished" podID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerID="0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f" exitCode=0 Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.758838 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4bmb" event={"ID":"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961","Type":"ContainerDied","Data":"0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f"} Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.759017 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4bmb" event={"ID":"d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961","Type":"ContainerDied","Data":"8d126dbf0e3e3df37c9bdf9b506c82eb3b1fc2aeeeaaa756c8d8e45c806974dc"} Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.759042 4726 scope.go:117] "RemoveContainer" containerID="0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.758941 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4bmb" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.810853 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w4bmb"] Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.813806 4726 scope.go:117] "RemoveContainer" containerID="217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.819226 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w4bmb"] Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.838560 4726 scope.go:117] "RemoveContainer" containerID="34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.884377 4726 scope.go:117] "RemoveContainer" containerID="0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f" Nov 23 21:25:09 crc kubenswrapper[4726]: E1123 21:25:09.884802 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f\": container with ID starting with 0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f not found: ID does not exist" containerID="0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.884835 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f"} err="failed to get container status \"0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f\": rpc error: code = NotFound desc = could not find container \"0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f\": container with ID starting with 0cc11fbafc9dd4cf89bbc555700ab2f47195218820f3faef62da564e139a0d9f not found: ID does not exist" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.884860 4726 scope.go:117] "RemoveContainer" containerID="217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc" Nov 23 21:25:09 crc kubenswrapper[4726]: E1123 21:25:09.885328 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc\": container with ID starting with 217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc not found: ID does not exist" containerID="217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.885352 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc"} err="failed to get container status \"217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc\": rpc error: code = NotFound desc = could not find container \"217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc\": container with ID starting with 217e4425656bd199e72b7a3972643af9b2675b1df5bb7223af0c98c6d2bb67fc not found: ID does not exist" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.885367 4726 scope.go:117] "RemoveContainer" containerID="34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2" Nov 23 21:25:09 crc kubenswrapper[4726]: E1123 21:25:09.886390 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2\": container with ID starting with 34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2 not found: ID does not exist" containerID="34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2" Nov 23 21:25:09 crc kubenswrapper[4726]: I1123 21:25:09.886486 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2"} err="failed to get container status \"34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2\": rpc error: code = NotFound desc = could not find container \"34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2\": container with ID starting with 34a21c657cce63d2c97595d3a1ce10e8c41d674a7e01b7dd44266e1ea8efe9d2 not found: ID does not exist" Nov 23 21:25:10 crc kubenswrapper[4726]: I1123 21:25:10.601409 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" path="/var/lib/kubelet/pods/d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961/volumes" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.104329 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gz4k6/must-gather-cw2lq"] Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.104626 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" podUID="d5709a4c-d883-4f50-b622-f482770107ff" containerName="copy" containerID="cri-o://aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070" gracePeriod=2 Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.112410 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gz4k6/must-gather-cw2lq"] Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.589096 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gz4k6_must-gather-cw2lq_d5709a4c-d883-4f50-b622-f482770107ff/copy/0.log" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.590064 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.744728 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f56pq\" (UniqueName: \"kubernetes.io/projected/d5709a4c-d883-4f50-b622-f482770107ff-kube-api-access-f56pq\") pod \"d5709a4c-d883-4f50-b622-f482770107ff\" (UID: \"d5709a4c-d883-4f50-b622-f482770107ff\") " Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.744786 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d5709a4c-d883-4f50-b622-f482770107ff-must-gather-output\") pod \"d5709a4c-d883-4f50-b622-f482770107ff\" (UID: \"d5709a4c-d883-4f50-b622-f482770107ff\") " Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.766137 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5709a4c-d883-4f50-b622-f482770107ff-kube-api-access-f56pq" (OuterVolumeSpecName: "kube-api-access-f56pq") pod "d5709a4c-d883-4f50-b622-f482770107ff" (UID: "d5709a4c-d883-4f50-b622-f482770107ff"). InnerVolumeSpecName "kube-api-access-f56pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.786028 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gz4k6_must-gather-cw2lq_d5709a4c-d883-4f50-b622-f482770107ff/copy/0.log" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.786353 4726 generic.go:334] "Generic (PLEG): container finished" podID="d5709a4c-d883-4f50-b622-f482770107ff" containerID="aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070" exitCode=143 Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.786415 4726 scope.go:117] "RemoveContainer" containerID="aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.786622 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gz4k6/must-gather-cw2lq" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.823194 4726 scope.go:117] "RemoveContainer" containerID="4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.847863 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f56pq\" (UniqueName: \"kubernetes.io/projected/d5709a4c-d883-4f50-b622-f482770107ff-kube-api-access-f56pq\") on node \"crc\" DevicePath \"\"" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.879513 4726 scope.go:117] "RemoveContainer" containerID="aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070" Nov 23 21:25:11 crc kubenswrapper[4726]: E1123 21:25:11.884996 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070\": container with ID starting with aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070 not found: ID does not exist" containerID="aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.885042 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070"} err="failed to get container status \"aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070\": rpc error: code = NotFound desc = could not find container \"aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070\": container with ID starting with aa8e92a325adc4bfbded28454d8de7f30ec094111e9b696510172b56b312e070 not found: ID does not exist" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.885068 4726 scope.go:117] "RemoveContainer" containerID="4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce" Nov 23 21:25:11 crc kubenswrapper[4726]: E1123 21:25:11.885965 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce\": container with ID starting with 4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce not found: ID does not exist" containerID="4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.885988 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce"} err="failed to get container status \"4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce\": rpc error: code = NotFound desc = could not find container \"4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce\": container with ID starting with 4306491fe1d7781c4539c23b0cc96edc6c4af04be199ab9ec82bf946ffdc00ce not found: ID does not exist" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.942623 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5709a4c-d883-4f50-b622-f482770107ff-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d5709a4c-d883-4f50-b622-f482770107ff" (UID: "d5709a4c-d883-4f50-b622-f482770107ff"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:25:11 crc kubenswrapper[4726]: I1123 21:25:11.953838 4726 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d5709a4c-d883-4f50-b622-f482770107ff-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 23 21:25:12 crc kubenswrapper[4726]: I1123 21:25:12.601997 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5709a4c-d883-4f50-b622-f482770107ff" path="/var/lib/kubelet/pods/d5709a4c-d883-4f50-b622-f482770107ff/volumes" Nov 23 21:25:14 crc kubenswrapper[4726]: I1123 21:25:14.779659 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j9vwx" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="registry-server" probeResult="failure" output=< Nov 23 21:25:14 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 21:25:14 crc kubenswrapper[4726]: > Nov 23 21:25:23 crc kubenswrapper[4726]: I1123 21:25:23.770587 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:25:23 crc kubenswrapper[4726]: I1123 21:25:23.824566 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:25:24 crc kubenswrapper[4726]: I1123 21:25:24.012330 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9vwx"] Nov 23 21:25:24 crc kubenswrapper[4726]: I1123 21:25:24.921427 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j9vwx" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="registry-server" containerID="cri-o://35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2" gracePeriod=2 Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.360631 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.460989 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-catalog-content\") pod \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.461456 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-utilities\") pod \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.461535 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxfgz\" (UniqueName: \"kubernetes.io/projected/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-kube-api-access-cxfgz\") pod \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\" (UID: \"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef\") " Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.463380 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-utilities" (OuterVolumeSpecName: "utilities") pod "bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" (UID: "bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.474927 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-kube-api-access-cxfgz" (OuterVolumeSpecName: "kube-api-access-cxfgz") pod "bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" (UID: "bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef"). InnerVolumeSpecName "kube-api-access-cxfgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.563818 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.563898 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxfgz\" (UniqueName: \"kubernetes.io/projected/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-kube-api-access-cxfgz\") on node \"crc\" DevicePath \"\"" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.583511 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" (UID: "bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.665623 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.932818 4726 generic.go:334] "Generic (PLEG): container finished" podID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerID="35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2" exitCode=0 Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.932880 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9vwx" event={"ID":"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef","Type":"ContainerDied","Data":"35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2"} Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.932915 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9vwx" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.932928 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9vwx" event={"ID":"bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef","Type":"ContainerDied","Data":"31678dbb77fc30cfc6776448939c3de51d8fad4eea51dcdaa10f746531b8a4a8"} Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.932948 4726 scope.go:117] "RemoveContainer" containerID="35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.956455 4726 scope.go:117] "RemoveContainer" containerID="0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.976999 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9vwx"] Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.983104 4726 scope.go:117] "RemoveContainer" containerID="a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9" Nov 23 21:25:25 crc kubenswrapper[4726]: I1123 21:25:25.985925 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j9vwx"] Nov 23 21:25:26 crc kubenswrapper[4726]: I1123 21:25:26.021016 4726 scope.go:117] "RemoveContainer" containerID="35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2" Nov 23 21:25:26 crc kubenswrapper[4726]: E1123 21:25:26.021531 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2\": container with ID starting with 35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2 not found: ID does not exist" containerID="35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2" Nov 23 21:25:26 crc kubenswrapper[4726]: I1123 21:25:26.021562 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2"} err="failed to get container status \"35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2\": rpc error: code = NotFound desc = could not find container \"35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2\": container with ID starting with 35b37e2dafd95e897b2f615fc40cdea4e3f6acc5fcec913e6381c3353233d6b2 not found: ID does not exist" Nov 23 21:25:26 crc kubenswrapper[4726]: I1123 21:25:26.021582 4726 scope.go:117] "RemoveContainer" containerID="0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82" Nov 23 21:25:26 crc kubenswrapper[4726]: E1123 21:25:26.022217 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82\": container with ID starting with 0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82 not found: ID does not exist" containerID="0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82" Nov 23 21:25:26 crc kubenswrapper[4726]: I1123 21:25:26.022243 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82"} err="failed to get container status \"0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82\": rpc error: code = NotFound desc = could not find container \"0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82\": container with ID starting with 0665f58abdd4f27e8bd370c63149302e99db44860420e6813d129b487a898a82 not found: ID does not exist" Nov 23 21:25:26 crc kubenswrapper[4726]: I1123 21:25:26.022256 4726 scope.go:117] "RemoveContainer" containerID="a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9" Nov 23 21:25:26 crc kubenswrapper[4726]: E1123 21:25:26.022565 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9\": container with ID starting with a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9 not found: ID does not exist" containerID="a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9" Nov 23 21:25:26 crc kubenswrapper[4726]: I1123 21:25:26.022591 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9"} err="failed to get container status \"a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9\": rpc error: code = NotFound desc = could not find container \"a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9\": container with ID starting with a4e1695312f17ef35c70a7c93fd7348ed5955b16cf19b67ad1777a59b26e49c9 not found: ID does not exist" Nov 23 21:25:26 crc kubenswrapper[4726]: I1123 21:25:26.610203 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" path="/var/lib/kubelet/pods/bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef/volumes" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.127209 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bfm2h"] Nov 23 21:25:57 crc kubenswrapper[4726]: E1123 21:25:57.128326 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="extract-content" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128346 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="extract-content" Nov 23 21:25:57 crc kubenswrapper[4726]: E1123 21:25:57.128405 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5709a4c-d883-4f50-b622-f482770107ff" containerName="gather" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128414 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5709a4c-d883-4f50-b622-f482770107ff" containerName="gather" Nov 23 21:25:57 crc kubenswrapper[4726]: E1123 21:25:57.128430 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5709a4c-d883-4f50-b622-f482770107ff" containerName="copy" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128437 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5709a4c-d883-4f50-b622-f482770107ff" containerName="copy" Nov 23 21:25:57 crc kubenswrapper[4726]: E1123 21:25:57.128461 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerName="extract-content" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128468 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerName="extract-content" Nov 23 21:25:57 crc kubenswrapper[4726]: E1123 21:25:57.128488 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerName="extract-utilities" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128499 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerName="extract-utilities" Nov 23 21:25:57 crc kubenswrapper[4726]: E1123 21:25:57.128510 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="registry-server" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128518 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="registry-server" Nov 23 21:25:57 crc kubenswrapper[4726]: E1123 21:25:57.128542 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerName="registry-server" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128550 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerName="registry-server" Nov 23 21:25:57 crc kubenswrapper[4726]: E1123 21:25:57.128566 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="extract-utilities" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128573 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="extract-utilities" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128791 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc59ef08-f1a9-4b5b-8f88-05f0ee1c46ef" containerName="registry-server" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128820 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5709a4c-d883-4f50-b622-f482770107ff" containerName="gather" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128833 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7bdf9a6-ec8f-41a4-9b3e-bce8b0d64961" containerName="registry-server" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.128851 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5709a4c-d883-4f50-b622-f482770107ff" containerName="copy" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.130572 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.139540 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfm2h"] Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.213562 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-utilities\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.213854 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-catalog-content\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.214123 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msxxs\" (UniqueName: \"kubernetes.io/projected/b42393c2-4342-4f36-9273-2c6b83ba0338-kube-api-access-msxxs\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.315565 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msxxs\" (UniqueName: \"kubernetes.io/projected/b42393c2-4342-4f36-9273-2c6b83ba0338-kube-api-access-msxxs\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.315852 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-utilities\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.316006 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-catalog-content\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.316455 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-utilities\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.316546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-catalog-content\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.333780 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msxxs\" (UniqueName: \"kubernetes.io/projected/b42393c2-4342-4f36-9273-2c6b83ba0338-kube-api-access-msxxs\") pod \"redhat-marketplace-bfm2h\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:57 crc kubenswrapper[4726]: I1123 21:25:57.464731 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:25:58 crc kubenswrapper[4726]: I1123 21:25:58.014299 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfm2h"] Nov 23 21:25:58 crc kubenswrapper[4726]: I1123 21:25:58.248858 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfm2h" event={"ID":"b42393c2-4342-4f36-9273-2c6b83ba0338","Type":"ContainerStarted","Data":"801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c"} Nov 23 21:25:58 crc kubenswrapper[4726]: I1123 21:25:58.248939 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfm2h" event={"ID":"b42393c2-4342-4f36-9273-2c6b83ba0338","Type":"ContainerStarted","Data":"38dc266962da5fc9725017bf50ce2798c74ce2593d1b2271d35b8eb4be81029a"} Nov 23 21:25:59 crc kubenswrapper[4726]: I1123 21:25:59.259330 4726 generic.go:334] "Generic (PLEG): container finished" podID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerID="801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c" exitCode=0 Nov 23 21:25:59 crc kubenswrapper[4726]: I1123 21:25:59.259497 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfm2h" event={"ID":"b42393c2-4342-4f36-9273-2c6b83ba0338","Type":"ContainerDied","Data":"801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c"} Nov 23 21:26:00 crc kubenswrapper[4726]: I1123 21:26:00.275859 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfm2h" event={"ID":"b42393c2-4342-4f36-9273-2c6b83ba0338","Type":"ContainerStarted","Data":"0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb"} Nov 23 21:26:01 crc kubenswrapper[4726]: I1123 21:26:01.290966 4726 generic.go:334] "Generic (PLEG): container finished" podID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerID="0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb" exitCode=0 Nov 23 21:26:01 crc kubenswrapper[4726]: I1123 21:26:01.291095 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfm2h" event={"ID":"b42393c2-4342-4f36-9273-2c6b83ba0338","Type":"ContainerDied","Data":"0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb"} Nov 23 21:26:02 crc kubenswrapper[4726]: I1123 21:26:02.301354 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfm2h" event={"ID":"b42393c2-4342-4f36-9273-2c6b83ba0338","Type":"ContainerStarted","Data":"114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc"} Nov 23 21:26:02 crc kubenswrapper[4726]: I1123 21:26:02.331782 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bfm2h" podStartSLOduration=2.8233952909999998 podStartE2EDuration="5.331757994s" podCreationTimestamp="2025-11-23 21:25:57 +0000 UTC" firstStartedPulling="2025-11-23 21:25:59.262493551 +0000 UTC m=+4667.411534517" lastFinishedPulling="2025-11-23 21:26:01.770856254 +0000 UTC m=+4669.919897220" observedRunningTime="2025-11-23 21:26:02.324731738 +0000 UTC m=+4670.473772704" watchObservedRunningTime="2025-11-23 21:26:02.331757994 +0000 UTC m=+4670.480798950" Nov 23 21:26:07 crc kubenswrapper[4726]: I1123 21:26:07.464957 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:26:07 crc kubenswrapper[4726]: I1123 21:26:07.465487 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:26:07 crc kubenswrapper[4726]: I1123 21:26:07.532787 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:26:08 crc kubenswrapper[4726]: I1123 21:26:08.409961 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:26:08 crc kubenswrapper[4726]: I1123 21:26:08.458736 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfm2h"] Nov 23 21:26:10 crc kubenswrapper[4726]: I1123 21:26:10.384459 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bfm2h" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerName="registry-server" containerID="cri-o://114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc" gracePeriod=2 Nov 23 21:26:10 crc kubenswrapper[4726]: I1123 21:26:10.834644 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:26:10 crc kubenswrapper[4726]: I1123 21:26:10.970602 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-utilities\") pod \"b42393c2-4342-4f36-9273-2c6b83ba0338\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " Nov 23 21:26:10 crc kubenswrapper[4726]: I1123 21:26:10.970707 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msxxs\" (UniqueName: \"kubernetes.io/projected/b42393c2-4342-4f36-9273-2c6b83ba0338-kube-api-access-msxxs\") pod \"b42393c2-4342-4f36-9273-2c6b83ba0338\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " Nov 23 21:26:10 crc kubenswrapper[4726]: I1123 21:26:10.970761 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-catalog-content\") pod \"b42393c2-4342-4f36-9273-2c6b83ba0338\" (UID: \"b42393c2-4342-4f36-9273-2c6b83ba0338\") " Nov 23 21:26:10 crc kubenswrapper[4726]: I1123 21:26:10.971884 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-utilities" (OuterVolumeSpecName: "utilities") pod "b42393c2-4342-4f36-9273-2c6b83ba0338" (UID: "b42393c2-4342-4f36-9273-2c6b83ba0338"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:26:10 crc kubenswrapper[4726]: I1123 21:26:10.975556 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b42393c2-4342-4f36-9273-2c6b83ba0338-kube-api-access-msxxs" (OuterVolumeSpecName: "kube-api-access-msxxs") pod "b42393c2-4342-4f36-9273-2c6b83ba0338" (UID: "b42393c2-4342-4f36-9273-2c6b83ba0338"). InnerVolumeSpecName "kube-api-access-msxxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:26:10 crc kubenswrapper[4726]: I1123 21:26:10.994126 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b42393c2-4342-4f36-9273-2c6b83ba0338" (UID: "b42393c2-4342-4f36-9273-2c6b83ba0338"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.074705 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.074742 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b42393c2-4342-4f36-9273-2c6b83ba0338-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.074756 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msxxs\" (UniqueName: \"kubernetes.io/projected/b42393c2-4342-4f36-9273-2c6b83ba0338-kube-api-access-msxxs\") on node \"crc\" DevicePath \"\"" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.398481 4726 generic.go:334] "Generic (PLEG): container finished" podID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerID="114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc" exitCode=0 Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.398614 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfm2h" event={"ID":"b42393c2-4342-4f36-9273-2c6b83ba0338","Type":"ContainerDied","Data":"114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc"} Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.398665 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfm2h" event={"ID":"b42393c2-4342-4f36-9273-2c6b83ba0338","Type":"ContainerDied","Data":"38dc266962da5fc9725017bf50ce2798c74ce2593d1b2271d35b8eb4be81029a"} Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.398742 4726 scope.go:117] "RemoveContainer" containerID="114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.399249 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfm2h" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.428443 4726 scope.go:117] "RemoveContainer" containerID="0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.447610 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfm2h"] Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.465706 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfm2h"] Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.476190 4726 scope.go:117] "RemoveContainer" containerID="801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.518731 4726 scope.go:117] "RemoveContainer" containerID="114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc" Nov 23 21:26:11 crc kubenswrapper[4726]: E1123 21:26:11.519276 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc\": container with ID starting with 114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc not found: ID does not exist" containerID="114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.519317 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc"} err="failed to get container status \"114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc\": rpc error: code = NotFound desc = could not find container \"114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc\": container with ID starting with 114a84f6b2b5880d2473b0bf5bd393d55a024e3eee8f5b14a5f1b8f4657062cc not found: ID does not exist" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.519344 4726 scope.go:117] "RemoveContainer" containerID="0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb" Nov 23 21:26:11 crc kubenswrapper[4726]: E1123 21:26:11.519752 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb\": container with ID starting with 0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb not found: ID does not exist" containerID="0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.519784 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb"} err="failed to get container status \"0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb\": rpc error: code = NotFound desc = could not find container \"0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb\": container with ID starting with 0887bdc08af2bd2122897ffe14ddc8c9bd8c4a6429898c4f904d3d2018a040bb not found: ID does not exist" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.519804 4726 scope.go:117] "RemoveContainer" containerID="801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c" Nov 23 21:26:11 crc kubenswrapper[4726]: E1123 21:26:11.520200 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c\": container with ID starting with 801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c not found: ID does not exist" containerID="801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c" Nov 23 21:26:11 crc kubenswrapper[4726]: I1123 21:26:11.520247 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c"} err="failed to get container status \"801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c\": rpc error: code = NotFound desc = could not find container \"801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c\": container with ID starting with 801d723cf12a3c2866a18e3b66774ebb6996b9c27ed56735162672acfb3a410c not found: ID does not exist" Nov 23 21:26:12 crc kubenswrapper[4726]: I1123 21:26:12.608481 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" path="/var/lib/kubelet/pods/b42393c2-4342-4f36-9273-2c6b83ba0338/volumes" Nov 23 21:26:21 crc kubenswrapper[4726]: I1123 21:26:21.784250 4726 scope.go:117] "RemoveContainer" containerID="81d5f91df0f17432bdea6edbba2d11a0ad818aa014306a3f300498a14633e360" Nov 23 21:26:39 crc kubenswrapper[4726]: I1123 21:26:39.043180 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:26:39 crc kubenswrapper[4726]: I1123 21:26:39.043846 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:27:09 crc kubenswrapper[4726]: I1123 21:27:09.042985 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:27:09 crc kubenswrapper[4726]: I1123 21:27:09.043597 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:27:39 crc kubenswrapper[4726]: I1123 21:27:39.043217 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:27:39 crc kubenswrapper[4726]: I1123 21:27:39.043964 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:27:39 crc kubenswrapper[4726]: I1123 21:27:39.044021 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 21:27:39 crc kubenswrapper[4726]: I1123 21:27:39.044716 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d78047930e4aba8b59f1ecbdc499155597ea2347de6f9220fe2147928a3ef4b3"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 21:27:39 crc kubenswrapper[4726]: I1123 21:27:39.044795 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://d78047930e4aba8b59f1ecbdc499155597ea2347de6f9220fe2147928a3ef4b3" gracePeriod=600 Nov 23 21:27:39 crc kubenswrapper[4726]: I1123 21:27:39.346557 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="d78047930e4aba8b59f1ecbdc499155597ea2347de6f9220fe2147928a3ef4b3" exitCode=0 Nov 23 21:27:39 crc kubenswrapper[4726]: I1123 21:27:39.346681 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"d78047930e4aba8b59f1ecbdc499155597ea2347de6f9220fe2147928a3ef4b3"} Nov 23 21:27:39 crc kubenswrapper[4726]: I1123 21:27:39.347077 4726 scope.go:117] "RemoveContainer" containerID="ac9588a9541742c3589e90fb12c2a876081d8916c49c6f4ece7f05520cc3c51c" Nov 23 21:27:40 crc kubenswrapper[4726]: I1123 21:27:40.367491 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d"} Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.196415 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rtcdk/must-gather-xkxgs"] Nov 23 21:27:48 crc kubenswrapper[4726]: E1123 21:27:48.197511 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerName="extract-content" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.197526 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerName="extract-content" Nov 23 21:27:48 crc kubenswrapper[4726]: E1123 21:27:48.197551 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerName="registry-server" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.197558 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerName="registry-server" Nov 23 21:27:48 crc kubenswrapper[4726]: E1123 21:27:48.197587 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerName="extract-utilities" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.197595 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerName="extract-utilities" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.197814 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b42393c2-4342-4f36-9273-2c6b83ba0338" containerName="registry-server" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.200061 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.209626 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rtcdk"/"kube-root-ca.crt" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.209638 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rtcdk"/"openshift-service-ca.crt" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.211136 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rtcdk"/"default-dockercfg-sdtsd" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.223374 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rtcdk/must-gather-xkxgs"] Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.253491 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gthfx\" (UniqueName: \"kubernetes.io/projected/637adc9a-b276-423b-a8ed-bf6f3143ec97-kube-api-access-gthfx\") pod \"must-gather-xkxgs\" (UID: \"637adc9a-b276-423b-a8ed-bf6f3143ec97\") " pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.253580 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/637adc9a-b276-423b-a8ed-bf6f3143ec97-must-gather-output\") pod \"must-gather-xkxgs\" (UID: \"637adc9a-b276-423b-a8ed-bf6f3143ec97\") " pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.356128 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gthfx\" (UniqueName: \"kubernetes.io/projected/637adc9a-b276-423b-a8ed-bf6f3143ec97-kube-api-access-gthfx\") pod \"must-gather-xkxgs\" (UID: \"637adc9a-b276-423b-a8ed-bf6f3143ec97\") " pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.356487 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/637adc9a-b276-423b-a8ed-bf6f3143ec97-must-gather-output\") pod \"must-gather-xkxgs\" (UID: \"637adc9a-b276-423b-a8ed-bf6f3143ec97\") " pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.356955 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/637adc9a-b276-423b-a8ed-bf6f3143ec97-must-gather-output\") pod \"must-gather-xkxgs\" (UID: \"637adc9a-b276-423b-a8ed-bf6f3143ec97\") " pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.379861 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gthfx\" (UniqueName: \"kubernetes.io/projected/637adc9a-b276-423b-a8ed-bf6f3143ec97-kube-api-access-gthfx\") pod \"must-gather-xkxgs\" (UID: \"637adc9a-b276-423b-a8ed-bf6f3143ec97\") " pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:27:48 crc kubenswrapper[4726]: I1123 21:27:48.524164 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:27:49 crc kubenswrapper[4726]: I1123 21:27:49.040546 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rtcdk/must-gather-xkxgs"] Nov 23 21:27:49 crc kubenswrapper[4726]: I1123 21:27:49.467317 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" event={"ID":"637adc9a-b276-423b-a8ed-bf6f3143ec97","Type":"ContainerStarted","Data":"cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63"} Nov 23 21:27:49 crc kubenswrapper[4726]: I1123 21:27:49.467692 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" event={"ID":"637adc9a-b276-423b-a8ed-bf6f3143ec97","Type":"ContainerStarted","Data":"2bfdac50d74bc21dee554cfbe04480ab1c0d5435a8bf05b43d8ec2756d771a5f"} Nov 23 21:27:50 crc kubenswrapper[4726]: I1123 21:27:50.479390 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" event={"ID":"637adc9a-b276-423b-a8ed-bf6f3143ec97","Type":"ContainerStarted","Data":"f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21"} Nov 23 21:27:50 crc kubenswrapper[4726]: I1123 21:27:50.505545 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" podStartSLOduration=2.5055252 podStartE2EDuration="2.5055252s" podCreationTimestamp="2025-11-23 21:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:27:50.496094028 +0000 UTC m=+4778.645134994" watchObservedRunningTime="2025-11-23 21:27:50.5055252 +0000 UTC m=+4778.654566176" Nov 23 21:27:51 crc kubenswrapper[4726]: E1123 21:27:51.388580 4726 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.110:46346->38.102.83.110:43095: write tcp 38.102.83.110:46346->38.102.83.110:43095: write: connection reset by peer Nov 23 21:27:53 crc kubenswrapper[4726]: I1123 21:27:53.806621 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-xkvff"] Nov 23 21:27:53 crc kubenswrapper[4726]: I1123 21:27:53.808367 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:27:53 crc kubenswrapper[4726]: I1123 21:27:53.980984 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mjqp\" (UniqueName: \"kubernetes.io/projected/611572db-c135-4e99-972f-f18831959362-kube-api-access-4mjqp\") pod \"crc-debug-xkvff\" (UID: \"611572db-c135-4e99-972f-f18831959362\") " pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:27:53 crc kubenswrapper[4726]: I1123 21:27:53.981048 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/611572db-c135-4e99-972f-f18831959362-host\") pod \"crc-debug-xkvff\" (UID: \"611572db-c135-4e99-972f-f18831959362\") " pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:27:54 crc kubenswrapper[4726]: I1123 21:27:54.082904 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mjqp\" (UniqueName: \"kubernetes.io/projected/611572db-c135-4e99-972f-f18831959362-kube-api-access-4mjqp\") pod \"crc-debug-xkvff\" (UID: \"611572db-c135-4e99-972f-f18831959362\") " pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:27:54 crc kubenswrapper[4726]: I1123 21:27:54.082963 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/611572db-c135-4e99-972f-f18831959362-host\") pod \"crc-debug-xkvff\" (UID: \"611572db-c135-4e99-972f-f18831959362\") " pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:27:54 crc kubenswrapper[4726]: I1123 21:27:54.083064 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/611572db-c135-4e99-972f-f18831959362-host\") pod \"crc-debug-xkvff\" (UID: \"611572db-c135-4e99-972f-f18831959362\") " pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:27:54 crc kubenswrapper[4726]: I1123 21:27:54.103569 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mjqp\" (UniqueName: \"kubernetes.io/projected/611572db-c135-4e99-972f-f18831959362-kube-api-access-4mjqp\") pod \"crc-debug-xkvff\" (UID: \"611572db-c135-4e99-972f-f18831959362\") " pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:27:54 crc kubenswrapper[4726]: I1123 21:27:54.124748 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:27:54 crc kubenswrapper[4726]: W1123 21:27:54.149968 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod611572db_c135_4e99_972f_f18831959362.slice/crio-730c18d3c2edbd593ed54b4fefce9a77a23dc0bdc02f6c977022b0b4be267f1d WatchSource:0}: Error finding container 730c18d3c2edbd593ed54b4fefce9a77a23dc0bdc02f6c977022b0b4be267f1d: Status 404 returned error can't find the container with id 730c18d3c2edbd593ed54b4fefce9a77a23dc0bdc02f6c977022b0b4be267f1d Nov 23 21:27:54 crc kubenswrapper[4726]: I1123 21:27:54.531639 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/crc-debug-xkvff" event={"ID":"611572db-c135-4e99-972f-f18831959362","Type":"ContainerStarted","Data":"b91f3a03df3fd8d1126fa4d6227835f7bac21c75b99470d5c634f9587b8806f9"} Nov 23 21:27:54 crc kubenswrapper[4726]: I1123 21:27:54.532068 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/crc-debug-xkvff" event={"ID":"611572db-c135-4e99-972f-f18831959362","Type":"ContainerStarted","Data":"730c18d3c2edbd593ed54b4fefce9a77a23dc0bdc02f6c977022b0b4be267f1d"} Nov 23 21:27:54 crc kubenswrapper[4726]: I1123 21:27:54.553496 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rtcdk/crc-debug-xkvff" podStartSLOduration=1.553479944 podStartE2EDuration="1.553479944s" podCreationTimestamp="2025-11-23 21:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:27:54.549142353 +0000 UTC m=+4782.698183309" watchObservedRunningTime="2025-11-23 21:27:54.553479944 +0000 UTC m=+4782.702520900" Nov 23 21:28:27 crc kubenswrapper[4726]: I1123 21:28:27.233124 4726 generic.go:334] "Generic (PLEG): container finished" podID="611572db-c135-4e99-972f-f18831959362" containerID="b91f3a03df3fd8d1126fa4d6227835f7bac21c75b99470d5c634f9587b8806f9" exitCode=0 Nov 23 21:28:27 crc kubenswrapper[4726]: I1123 21:28:27.233287 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/crc-debug-xkvff" event={"ID":"611572db-c135-4e99-972f-f18831959362","Type":"ContainerDied","Data":"b91f3a03df3fd8d1126fa4d6227835f7bac21c75b99470d5c634f9587b8806f9"} Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.354210 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.383174 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-xkvff"] Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.391673 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-xkvff"] Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.505523 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/611572db-c135-4e99-972f-f18831959362-host\") pod \"611572db-c135-4e99-972f-f18831959362\" (UID: \"611572db-c135-4e99-972f-f18831959362\") " Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.505580 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mjqp\" (UniqueName: \"kubernetes.io/projected/611572db-c135-4e99-972f-f18831959362-kube-api-access-4mjqp\") pod \"611572db-c135-4e99-972f-f18831959362\" (UID: \"611572db-c135-4e99-972f-f18831959362\") " Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.506704 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/611572db-c135-4e99-972f-f18831959362-host" (OuterVolumeSpecName: "host") pod "611572db-c135-4e99-972f-f18831959362" (UID: "611572db-c135-4e99-972f-f18831959362"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.511161 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/611572db-c135-4e99-972f-f18831959362-kube-api-access-4mjqp" (OuterVolumeSpecName: "kube-api-access-4mjqp") pod "611572db-c135-4e99-972f-f18831959362" (UID: "611572db-c135-4e99-972f-f18831959362"). InnerVolumeSpecName "kube-api-access-4mjqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.598136 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="611572db-c135-4e99-972f-f18831959362" path="/var/lib/kubelet/pods/611572db-c135-4e99-972f-f18831959362/volumes" Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.607649 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/611572db-c135-4e99-972f-f18831959362-host\") on node \"crc\" DevicePath \"\"" Nov 23 21:28:28 crc kubenswrapper[4726]: I1123 21:28:28.607688 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mjqp\" (UniqueName: \"kubernetes.io/projected/611572db-c135-4e99-972f-f18831959362-kube-api-access-4mjqp\") on node \"crc\" DevicePath \"\"" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.249280 4726 scope.go:117] "RemoveContainer" containerID="b91f3a03df3fd8d1126fa4d6227835f7bac21c75b99470d5c634f9587b8806f9" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.249314 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-xkvff" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.602004 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-b5x4r"] Nov 23 21:28:29 crc kubenswrapper[4726]: E1123 21:28:29.602452 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611572db-c135-4e99-972f-f18831959362" containerName="container-00" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.602471 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="611572db-c135-4e99-972f-f18831959362" containerName="container-00" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.602740 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="611572db-c135-4e99-972f-f18831959362" containerName="container-00" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.603623 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.727628 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-host\") pod \"crc-debug-b5x4r\" (UID: \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\") " pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.727747 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8lxn\" (UniqueName: \"kubernetes.io/projected/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-kube-api-access-r8lxn\") pod \"crc-debug-b5x4r\" (UID: \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\") " pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.829829 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-host\") pod \"crc-debug-b5x4r\" (UID: \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\") " pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.829908 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8lxn\" (UniqueName: \"kubernetes.io/projected/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-kube-api-access-r8lxn\") pod \"crc-debug-b5x4r\" (UID: \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\") " pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.830085 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-host\") pod \"crc-debug-b5x4r\" (UID: \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\") " pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.858506 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8lxn\" (UniqueName: \"kubernetes.io/projected/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-kube-api-access-r8lxn\") pod \"crc-debug-b5x4r\" (UID: \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\") " pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:29 crc kubenswrapper[4726]: I1123 21:28:29.934789 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:30 crc kubenswrapper[4726]: I1123 21:28:30.260786 4726 generic.go:334] "Generic (PLEG): container finished" podID="20e106e3-0cab-4d7d-8325-d6c407d3ae3a" containerID="b17aa32910d2f26e29661bbfa10031a674c532f1dff9b6a68d57145721d3367e" exitCode=0 Nov 23 21:28:30 crc kubenswrapper[4726]: I1123 21:28:30.260827 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" event={"ID":"20e106e3-0cab-4d7d-8325-d6c407d3ae3a","Type":"ContainerDied","Data":"b17aa32910d2f26e29661bbfa10031a674c532f1dff9b6a68d57145721d3367e"} Nov 23 21:28:30 crc kubenswrapper[4726]: I1123 21:28:30.260904 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" event={"ID":"20e106e3-0cab-4d7d-8325-d6c407d3ae3a","Type":"ContainerStarted","Data":"79b6211840f17c6ba9011a83788062020a428e0bf4940ba7f2a375951e508e07"} Nov 23 21:28:30 crc kubenswrapper[4726]: I1123 21:28:30.675582 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-b5x4r"] Nov 23 21:28:30 crc kubenswrapper[4726]: I1123 21:28:30.684213 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-b5x4r"] Nov 23 21:28:31 crc kubenswrapper[4726]: I1123 21:28:31.770855 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:31 crc kubenswrapper[4726]: I1123 21:28:31.868415 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8lxn\" (UniqueName: \"kubernetes.io/projected/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-kube-api-access-r8lxn\") pod \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\" (UID: \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\") " Nov 23 21:28:31 crc kubenswrapper[4726]: I1123 21:28:31.869113 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-host\") pod \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\" (UID: \"20e106e3-0cab-4d7d-8325-d6c407d3ae3a\") " Nov 23 21:28:31 crc kubenswrapper[4726]: I1123 21:28:31.869273 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-host" (OuterVolumeSpecName: "host") pod "20e106e3-0cab-4d7d-8325-d6c407d3ae3a" (UID: "20e106e3-0cab-4d7d-8325-d6c407d3ae3a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:28:31 crc kubenswrapper[4726]: I1123 21:28:31.869737 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-host\") on node \"crc\" DevicePath \"\"" Nov 23 21:28:31 crc kubenswrapper[4726]: I1123 21:28:31.876177 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-kube-api-access-r8lxn" (OuterVolumeSpecName: "kube-api-access-r8lxn") pod "20e106e3-0cab-4d7d-8325-d6c407d3ae3a" (UID: "20e106e3-0cab-4d7d-8325-d6c407d3ae3a"). InnerVolumeSpecName "kube-api-access-r8lxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:28:31 crc kubenswrapper[4726]: I1123 21:28:31.971439 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8lxn\" (UniqueName: \"kubernetes.io/projected/20e106e3-0cab-4d7d-8325-d6c407d3ae3a-kube-api-access-r8lxn\") on node \"crc\" DevicePath \"\"" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.033843 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-rlnsb"] Nov 23 21:28:32 crc kubenswrapper[4726]: E1123 21:28:32.034326 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e106e3-0cab-4d7d-8325-d6c407d3ae3a" containerName="container-00" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.034348 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e106e3-0cab-4d7d-8325-d6c407d3ae3a" containerName="container-00" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.034531 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="20e106e3-0cab-4d7d-8325-d6c407d3ae3a" containerName="container-00" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.035173 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.176029 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpt25\" (UniqueName: \"kubernetes.io/projected/566d7a72-4252-47f6-9a4e-6e4999886250-kube-api-access-fpt25\") pod \"crc-debug-rlnsb\" (UID: \"566d7a72-4252-47f6-9a4e-6e4999886250\") " pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.176291 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/566d7a72-4252-47f6-9a4e-6e4999886250-host\") pod \"crc-debug-rlnsb\" (UID: \"566d7a72-4252-47f6-9a4e-6e4999886250\") " pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.277033 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79b6211840f17c6ba9011a83788062020a428e0bf4940ba7f2a375951e508e07" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.277467 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-b5x4r" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.277903 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/566d7a72-4252-47f6-9a4e-6e4999886250-host\") pod \"crc-debug-rlnsb\" (UID: \"566d7a72-4252-47f6-9a4e-6e4999886250\") " pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.277997 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpt25\" (UniqueName: \"kubernetes.io/projected/566d7a72-4252-47f6-9a4e-6e4999886250-kube-api-access-fpt25\") pod \"crc-debug-rlnsb\" (UID: \"566d7a72-4252-47f6-9a4e-6e4999886250\") " pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.278052 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/566d7a72-4252-47f6-9a4e-6e4999886250-host\") pod \"crc-debug-rlnsb\" (UID: \"566d7a72-4252-47f6-9a4e-6e4999886250\") " pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.300047 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpt25\" (UniqueName: \"kubernetes.io/projected/566d7a72-4252-47f6-9a4e-6e4999886250-kube-api-access-fpt25\") pod \"crc-debug-rlnsb\" (UID: \"566d7a72-4252-47f6-9a4e-6e4999886250\") " pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.352631 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:32 crc kubenswrapper[4726]: W1123 21:28:32.382155 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod566d7a72_4252_47f6_9a4e_6e4999886250.slice/crio-9b784c119065b239e966682cab5b6ed4b02accecec96f3b5666f7b5fc9a0f1ac WatchSource:0}: Error finding container 9b784c119065b239e966682cab5b6ed4b02accecec96f3b5666f7b5fc9a0f1ac: Status 404 returned error can't find the container with id 9b784c119065b239e966682cab5b6ed4b02accecec96f3b5666f7b5fc9a0f1ac Nov 23 21:28:32 crc kubenswrapper[4726]: I1123 21:28:32.600388 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20e106e3-0cab-4d7d-8325-d6c407d3ae3a" path="/var/lib/kubelet/pods/20e106e3-0cab-4d7d-8325-d6c407d3ae3a/volumes" Nov 23 21:28:33 crc kubenswrapper[4726]: I1123 21:28:33.288382 4726 generic.go:334] "Generic (PLEG): container finished" podID="566d7a72-4252-47f6-9a4e-6e4999886250" containerID="d7bb19e7a6bb92e0e5236bdf24ef054ad03c9bccfb8309424d7d688645739b58" exitCode=0 Nov 23 21:28:33 crc kubenswrapper[4726]: I1123 21:28:33.288484 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" event={"ID":"566d7a72-4252-47f6-9a4e-6e4999886250","Type":"ContainerDied","Data":"d7bb19e7a6bb92e0e5236bdf24ef054ad03c9bccfb8309424d7d688645739b58"} Nov 23 21:28:33 crc kubenswrapper[4726]: I1123 21:28:33.288648 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" event={"ID":"566d7a72-4252-47f6-9a4e-6e4999886250","Type":"ContainerStarted","Data":"9b784c119065b239e966682cab5b6ed4b02accecec96f3b5666f7b5fc9a0f1ac"} Nov 23 21:28:33 crc kubenswrapper[4726]: I1123 21:28:33.330507 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-rlnsb"] Nov 23 21:28:33 crc kubenswrapper[4726]: I1123 21:28:33.340173 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rtcdk/crc-debug-rlnsb"] Nov 23 21:28:34 crc kubenswrapper[4726]: I1123 21:28:34.397319 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:28:34 crc kubenswrapper[4726]: I1123 21:28:34.534640 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpt25\" (UniqueName: \"kubernetes.io/projected/566d7a72-4252-47f6-9a4e-6e4999886250-kube-api-access-fpt25\") pod \"566d7a72-4252-47f6-9a4e-6e4999886250\" (UID: \"566d7a72-4252-47f6-9a4e-6e4999886250\") " Nov 23 21:28:34 crc kubenswrapper[4726]: I1123 21:28:34.534815 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/566d7a72-4252-47f6-9a4e-6e4999886250-host\") pod \"566d7a72-4252-47f6-9a4e-6e4999886250\" (UID: \"566d7a72-4252-47f6-9a4e-6e4999886250\") " Nov 23 21:28:34 crc kubenswrapper[4726]: I1123 21:28:34.535246 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/566d7a72-4252-47f6-9a4e-6e4999886250-host" (OuterVolumeSpecName: "host") pod "566d7a72-4252-47f6-9a4e-6e4999886250" (UID: "566d7a72-4252-47f6-9a4e-6e4999886250"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 23 21:28:34 crc kubenswrapper[4726]: I1123 21:28:34.548168 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/566d7a72-4252-47f6-9a4e-6e4999886250-kube-api-access-fpt25" (OuterVolumeSpecName: "kube-api-access-fpt25") pod "566d7a72-4252-47f6-9a4e-6e4999886250" (UID: "566d7a72-4252-47f6-9a4e-6e4999886250"). InnerVolumeSpecName "kube-api-access-fpt25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:28:34 crc kubenswrapper[4726]: I1123 21:28:34.599649 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="566d7a72-4252-47f6-9a4e-6e4999886250" path="/var/lib/kubelet/pods/566d7a72-4252-47f6-9a4e-6e4999886250/volumes" Nov 23 21:28:34 crc kubenswrapper[4726]: I1123 21:28:34.636488 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/566d7a72-4252-47f6-9a4e-6e4999886250-host\") on node \"crc\" DevicePath \"\"" Nov 23 21:28:34 crc kubenswrapper[4726]: I1123 21:28:34.636523 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpt25\" (UniqueName: \"kubernetes.io/projected/566d7a72-4252-47f6-9a4e-6e4999886250-kube-api-access-fpt25\") on node \"crc\" DevicePath \"\"" Nov 23 21:28:35 crc kubenswrapper[4726]: I1123 21:28:35.306715 4726 scope.go:117] "RemoveContainer" containerID="d7bb19e7a6bb92e0e5236bdf24ef054ad03c9bccfb8309424d7d688645739b58" Nov 23 21:28:35 crc kubenswrapper[4726]: I1123 21:28:35.306737 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/crc-debug-rlnsb" Nov 23 21:29:39 crc kubenswrapper[4726]: I1123 21:29:39.043084 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:29:39 crc kubenswrapper[4726]: I1123 21:29:39.043567 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:29:47 crc kubenswrapper[4726]: I1123 21:29:47.224616 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-85855866bd-77958_cddea6c0-9834-4a47-b30c-ef32194ef6ea/barbican-api/0.log" Nov 23 21:29:47 crc kubenswrapper[4726]: I1123 21:29:47.280651 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-85855866bd-77958_cddea6c0-9834-4a47-b30c-ef32194ef6ea/barbican-api-log/0.log" Nov 23 21:29:47 crc kubenswrapper[4726]: I1123 21:29:47.437960 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-644b66b48-csfsq_02190dad-4d68-4cc1-937c-a45b260b4a95/barbican-keystone-listener/0.log" Nov 23 21:29:47 crc kubenswrapper[4726]: I1123 21:29:47.494340 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-644b66b48-csfsq_02190dad-4d68-4cc1-937c-a45b260b4a95/barbican-keystone-listener-log/0.log" Nov 23 21:29:47 crc kubenswrapper[4726]: I1123 21:29:47.713105 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b6bd669d5-fsxlx_3041645e-f5e5-4faf-b2de-365bcd156177/barbican-worker-log/0.log" Nov 23 21:29:47 crc kubenswrapper[4726]: I1123 21:29:47.720108 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b6bd669d5-fsxlx_3041645e-f5e5-4faf-b2de-365bcd156177/barbican-worker/0.log" Nov 23 21:29:47 crc kubenswrapper[4726]: I1123 21:29:47.904844 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-twdt7_e4c46a96-d9bb-4f6c-8fb8-2fa789ab67d3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:48 crc kubenswrapper[4726]: I1123 21:29:48.046719 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a3f087de-9790-4860-b3bd-1ab9e677d4bd/ceilometer-central-agent/0.log" Nov 23 21:29:48 crc kubenswrapper[4726]: I1123 21:29:48.085062 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a3f087de-9790-4860-b3bd-1ab9e677d4bd/ceilometer-notification-agent/0.log" Nov 23 21:29:48 crc kubenswrapper[4726]: I1123 21:29:48.131073 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a3f087de-9790-4860-b3bd-1ab9e677d4bd/proxy-httpd/0.log" Nov 23 21:29:48 crc kubenswrapper[4726]: I1123 21:29:48.265533 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a3f087de-9790-4860-b3bd-1ab9e677d4bd/sg-core/0.log" Nov 23 21:29:48 crc kubenswrapper[4726]: I1123 21:29:48.368819 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-xrc2h_00e5cdd7-83a9-468b-a56f-db77fe0da375/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:48 crc kubenswrapper[4726]: I1123 21:29:48.526551 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-59pz2_0c1ed863-f501-4aec-bafe-2d7d828b8d78/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:48 crc kubenswrapper[4726]: I1123 21:29:48.732133 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275/cinder-api/0.log" Nov 23 21:29:48 crc kubenswrapper[4726]: I1123 21:29:48.762338 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3d7e8de0-4ebb-4fcb-a224-5d1fbb6cb275/cinder-api-log/0.log" Nov 23 21:29:49 crc kubenswrapper[4726]: I1123 21:29:49.015685 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047/cinder-backup/0.log" Nov 23 21:29:49 crc kubenswrapper[4726]: I1123 21:29:49.169355 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_ee6cd5a0-6b49-4ec3-a18f-5530b5bcb047/probe/0.log" Nov 23 21:29:49 crc kubenswrapper[4726]: I1123 21:29:49.199591 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5462b4dd-3aaf-4511-b08f-4cfe23a07a0d/cinder-scheduler/0.log" Nov 23 21:29:49 crc kubenswrapper[4726]: I1123 21:29:49.303399 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5462b4dd-3aaf-4511-b08f-4cfe23a07a0d/probe/0.log" Nov 23 21:29:49 crc kubenswrapper[4726]: I1123 21:29:49.832706 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a937c3c1-2bbf-4cf8-b5ac-9b55755df0af/probe/0.log" Nov 23 21:29:49 crc kubenswrapper[4726]: I1123 21:29:49.911616 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a937c3c1-2bbf-4cf8-b5ac-9b55755df0af/cinder-volume/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.046986 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-rd42k_74d7e759-c823-49e1-9480-87d93552cd4c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.148026 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jfpbr_77123683-ed45-435f-b733-f5cc3822169c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.288026 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7ddf789dd7-wznxr_6c62393d-31e9-4a24-9976-ffcf175d1443/init/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.598117 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7ddf789dd7-wznxr_6c62393d-31e9-4a24-9976-ffcf175d1443/init/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.606208 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb/glance-httpd/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.743222 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7ddf789dd7-wznxr_6c62393d-31e9-4a24-9976-ffcf175d1443/dnsmasq-dns/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.824928 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e4b7bb46-74e2-4aa0-8924-c8e325ad8cfb/glance-log/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.947594 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_97a134ee-4835-43e2-82b3-e1f19fa691cc/glance-httpd/0.log" Nov 23 21:29:50 crc kubenswrapper[4726]: I1123 21:29:50.973331 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_97a134ee-4835-43e2-82b3-e1f19fa691cc/glance-log/0.log" Nov 23 21:29:51 crc kubenswrapper[4726]: I1123 21:29:51.725008 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xnl4d_01fcd196-1da1-4fb2-a900-f3960109b195/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:51 crc kubenswrapper[4726]: I1123 21:29:51.729386 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c65c799bb-97hbb_30f69dae-368c-4ccf-9d4b-e7f943d7e662/horizon/0.log" Nov 23 21:29:51 crc kubenswrapper[4726]: I1123 21:29:51.755411 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c65c799bb-97hbb_30f69dae-368c-4ccf-9d4b-e7f943d7e662/horizon-log/0.log" Nov 23 21:29:51 crc kubenswrapper[4726]: I1123 21:29:51.936978 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-972gs_541068f2-1c18-4bc7-8c86-51f768e97783/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:52 crc kubenswrapper[4726]: I1123 21:29:52.121917 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-b76988558-2489x_1071728a-1612-4dc4-9cbe-25c043998eb5/keystone-api/0.log" Nov 23 21:29:52 crc kubenswrapper[4726]: I1123 21:29:52.219083 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29398861-kdlcf_d2340c37-0f69-48f7-94c1-e417bb7844a5/keystone-cron/0.log" Nov 23 21:29:52 crc kubenswrapper[4726]: I1123 21:29:52.296044 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_71a1f62d-5ad3-450a-824a-bff237a8252e/kube-state-metrics/0.log" Nov 23 21:29:52 crc kubenswrapper[4726]: I1123 21:29:52.403171 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-m2l2n_4d68c7f9-9c3a-4f39-ac3e-1a34894e3e91/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:52 crc kubenswrapper[4726]: I1123 21:29:52.588145 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f82bda1a-eb9c-4e50-9700-85c8f6a55336/manila-api-log/0.log" Nov 23 21:29:52 crc kubenswrapper[4726]: I1123 21:29:52.731532 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f82bda1a-eb9c-4e50-9700-85c8f6a55336/manila-api/0.log" Nov 23 21:29:52 crc kubenswrapper[4726]: I1123 21:29:52.888836 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_78ba73f5-3651-4b4b-9f8c-58fb451a655f/probe/0.log" Nov 23 21:29:52 crc kubenswrapper[4726]: I1123 21:29:52.935372 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_78ba73f5-3651-4b4b-9f8c-58fb451a655f/manila-scheduler/0.log" Nov 23 21:29:53 crc kubenswrapper[4726]: I1123 21:29:53.162500 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_ccbab0a3-e42f-42fd-8adb-ed1bae00642c/manila-share/0.log" Nov 23 21:29:53 crc kubenswrapper[4726]: I1123 21:29:53.240468 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_ccbab0a3-e42f-42fd-8adb-ed1bae00642c/probe/0.log" Nov 23 21:29:53 crc kubenswrapper[4726]: I1123 21:29:53.617017 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57cf6d944c-r5qgs_6e680832-25f3-4328-bf10-35bd83ab52b9/neutron-api/0.log" Nov 23 21:29:53 crc kubenswrapper[4726]: I1123 21:29:53.686449 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-57cf6d944c-r5qgs_6e680832-25f3-4328-bf10-35bd83ab52b9/neutron-httpd/0.log" Nov 23 21:29:53 crc kubenswrapper[4726]: I1123 21:29:53.726663 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-r2cg4_d233adb5-8473-4545-b91a-56ec95775b67/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:54 crc kubenswrapper[4726]: I1123 21:29:54.425814 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_29188283-47f2-45d2-b6ab-d3cc9b97faff/nova-api-log/0.log" Nov 23 21:29:54 crc kubenswrapper[4726]: I1123 21:29:54.559461 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a327e21d-682a-4006-9ef1-b6a07276fa3d/nova-cell0-conductor-conductor/0.log" Nov 23 21:29:54 crc kubenswrapper[4726]: I1123 21:29:54.926502 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_5d687a16-417f-4c97-b745-beba9b9429ae/nova-cell1-conductor-conductor/0.log" Nov 23 21:29:55 crc kubenswrapper[4726]: I1123 21:29:55.062731 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_609ae5f6-90fc-432a-820a-ca996f78c1c7/nova-cell1-novncproxy-novncproxy/0.log" Nov 23 21:29:55 crc kubenswrapper[4726]: I1123 21:29:55.132921 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_29188283-47f2-45d2-b6ab-d3cc9b97faff/nova-api-api/0.log" Nov 23 21:29:55 crc kubenswrapper[4726]: I1123 21:29:55.266208 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-b8pkj_89cb9922-a339-44bc-b1be-e7ffa127cf02/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:55 crc kubenswrapper[4726]: I1123 21:29:55.481382 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7e68d296-e4d0-4462-9b15-1700400d38fa/nova-metadata-log/0.log" Nov 23 21:29:55 crc kubenswrapper[4726]: I1123 21:29:55.929966 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e88e0fb1-897d-454c-9a23-8187592296e5/mysql-bootstrap/0.log" Nov 23 21:29:55 crc kubenswrapper[4726]: I1123 21:29:55.970766 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e324e799-3ac3-4e70-a035-822dff0d5972/nova-scheduler-scheduler/0.log" Nov 23 21:29:56 crc kubenswrapper[4726]: I1123 21:29:56.092491 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e88e0fb1-897d-454c-9a23-8187592296e5/mysql-bootstrap/0.log" Nov 23 21:29:56 crc kubenswrapper[4726]: I1123 21:29:56.331833 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e88e0fb1-897d-454c-9a23-8187592296e5/galera/0.log" Nov 23 21:29:56 crc kubenswrapper[4726]: I1123 21:29:56.436230 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3647d686-8ad6-4bad-9ca6-ffbaba9bba97/mysql-bootstrap/0.log" Nov 23 21:29:56 crc kubenswrapper[4726]: I1123 21:29:56.632295 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3647d686-8ad6-4bad-9ca6-ffbaba9bba97/mysql-bootstrap/0.log" Nov 23 21:29:56 crc kubenswrapper[4726]: I1123 21:29:56.655780 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3647d686-8ad6-4bad-9ca6-ffbaba9bba97/galera/0.log" Nov 23 21:29:56 crc kubenswrapper[4726]: I1123 21:29:56.886321 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_bfa50ecd-0a29-44c8-b219-d2e5df3709d1/openstackclient/0.log" Nov 23 21:29:57 crc kubenswrapper[4726]: I1123 21:29:57.009975 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-657sc_4cb17e51-7e2f-4012-96fd-5f6bd424429c/ovn-controller/0.log" Nov 23 21:29:57 crc kubenswrapper[4726]: I1123 21:29:57.257328 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jnv7b_fbe231ca-6343-4afb-b11b-a74c26a01a2f/openstack-network-exporter/0.log" Nov 23 21:29:57 crc kubenswrapper[4726]: I1123 21:29:57.487308 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sq7pg_ac53da75-5263-4d97-aa49-b4e8b345c854/ovsdb-server-init/0.log" Nov 23 21:29:57 crc kubenswrapper[4726]: I1123 21:29:57.535560 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7e68d296-e4d0-4462-9b15-1700400d38fa/nova-metadata-metadata/0.log" Nov 23 21:29:57 crc kubenswrapper[4726]: I1123 21:29:57.738125 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sq7pg_ac53da75-5263-4d97-aa49-b4e8b345c854/ovsdb-server-init/0.log" Nov 23 21:29:57 crc kubenswrapper[4726]: I1123 21:29:57.789274 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sq7pg_ac53da75-5263-4d97-aa49-b4e8b345c854/ovs-vswitchd/0.log" Nov 23 21:29:57 crc kubenswrapper[4726]: I1123 21:29:57.830241 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sq7pg_ac53da75-5263-4d97-aa49-b4e8b345c854/ovsdb-server/0.log" Nov 23 21:29:58 crc kubenswrapper[4726]: I1123 21:29:58.381277 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-q6k62_09d671dd-9524-4d36-8db8-a178093fdc70/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:29:58 crc kubenswrapper[4726]: I1123 21:29:58.489520 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d5bc7b6-0493-40ff-808b-6d13bb31c8c0/openstack-network-exporter/0.log" Nov 23 21:29:58 crc kubenswrapper[4726]: I1123 21:29:58.493439 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d5bc7b6-0493-40ff-808b-6d13bb31c8c0/ovn-northd/0.log" Nov 23 21:29:58 crc kubenswrapper[4726]: I1123 21:29:58.694726 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ff67ad55-4c63-45f3-b6f5-12f941bdd2a7/openstack-network-exporter/0.log" Nov 23 21:29:58 crc kubenswrapper[4726]: I1123 21:29:58.822677 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ff67ad55-4c63-45f3-b6f5-12f941bdd2a7/ovsdbserver-nb/0.log" Nov 23 21:29:58 crc kubenswrapper[4726]: I1123 21:29:58.900773 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1cb13d69-d5a0-4a7b-929b-ed077760725a/openstack-network-exporter/0.log" Nov 23 21:29:58 crc kubenswrapper[4726]: I1123 21:29:58.989564 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1cb13d69-d5a0-4a7b-929b-ed077760725a/ovsdbserver-sb/0.log" Nov 23 21:29:59 crc kubenswrapper[4726]: I1123 21:29:59.738488 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-fbb56fbb6-7fzgk_71d3aaf5-7d1c-46ac-8a80-5ab758b260d4/placement-log/0.log" Nov 23 21:29:59 crc kubenswrapper[4726]: I1123 21:29:59.904051 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-fbb56fbb6-7fzgk_71d3aaf5-7d1c-46ac-8a80-5ab758b260d4/placement-api/0.log" Nov 23 21:29:59 crc kubenswrapper[4726]: I1123 21:29:59.912165 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0883d1c4-a52b-4221-a651-8ba8c5463dd6/setup-container/0.log" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.159285 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw"] Nov 23 21:30:00 crc kubenswrapper[4726]: E1123 21:30:00.164104 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566d7a72-4252-47f6-9a4e-6e4999886250" containerName="container-00" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.164140 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="566d7a72-4252-47f6-9a4e-6e4999886250" containerName="container-00" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.164535 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="566d7a72-4252-47f6-9a4e-6e4999886250" containerName="container-00" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.165725 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.193676 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw"] Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.224479 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.224805 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.253922 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft6pr\" (UniqueName: \"kubernetes.io/projected/f07a9332-f2e2-41c7-8707-ac62bd160114-kube-api-access-ft6pr\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.254081 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f07a9332-f2e2-41c7-8707-ac62bd160114-secret-volume\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.254143 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f07a9332-f2e2-41c7-8707-ac62bd160114-config-volume\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.357938 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f07a9332-f2e2-41c7-8707-ac62bd160114-secret-volume\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.358036 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f07a9332-f2e2-41c7-8707-ac62bd160114-config-volume\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.358359 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft6pr\" (UniqueName: \"kubernetes.io/projected/f07a9332-f2e2-41c7-8707-ac62bd160114-kube-api-access-ft6pr\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.359184 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f07a9332-f2e2-41c7-8707-ac62bd160114-config-volume\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.378161 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f07a9332-f2e2-41c7-8707-ac62bd160114-secret-volume\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.383580 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft6pr\" (UniqueName: \"kubernetes.io/projected/f07a9332-f2e2-41c7-8707-ac62bd160114-kube-api-access-ft6pr\") pod \"collect-profiles-29398890-8pjcw\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.464785 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0883d1c4-a52b-4221-a651-8ba8c5463dd6/setup-container/0.log" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.464881 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f962af44-2b42-4924-81b8-872c96335906/setup-container/0.log" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.555916 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:00 crc kubenswrapper[4726]: I1123 21:30:00.558168 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0883d1c4-a52b-4221-a651-8ba8c5463dd6/rabbitmq/0.log" Nov 23 21:30:01 crc kubenswrapper[4726]: I1123 21:30:01.452931 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f962af44-2b42-4924-81b8-872c96335906/rabbitmq/0.log" Nov 23 21:30:01 crc kubenswrapper[4726]: I1123 21:30:01.485273 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-6rxz6_e1af3b69-28d2-4a52-8f13-7cb3e9831fb4/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:30:01 crc kubenswrapper[4726]: I1123 21:30:01.628563 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f962af44-2b42-4924-81b8-872c96335906/setup-container/0.log" Nov 23 21:30:01 crc kubenswrapper[4726]: I1123 21:30:01.666229 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw"] Nov 23 21:30:01 crc kubenswrapper[4726]: I1123 21:30:01.933350 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-mvddb_dc981c46-b927-4734-a0d1-061826f92146/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:30:01 crc kubenswrapper[4726]: I1123 21:30:01.956060 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pvpj5_275bc578-25bb-4340-9e5c-f2c296713110/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:30:02 crc kubenswrapper[4726]: I1123 21:30:02.152658 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" event={"ID":"f07a9332-f2e2-41c7-8707-ac62bd160114","Type":"ContainerStarted","Data":"0fc593c5e3c575671e09a72cdb4654986643642d745080b8502c0b5cfa91c8f1"} Nov 23 21:30:02 crc kubenswrapper[4726]: I1123 21:30:02.152705 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" event={"ID":"f07a9332-f2e2-41c7-8707-ac62bd160114","Type":"ContainerStarted","Data":"e43feee03109e5627964ce8a8ae4a21a863f38f11b9bb14093936485a7ecfdb5"} Nov 23 21:30:02 crc kubenswrapper[4726]: I1123 21:30:02.177064 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" podStartSLOduration=2.177038113 podStartE2EDuration="2.177038113s" podCreationTimestamp="2025-11-23 21:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-23 21:30:02.170313636 +0000 UTC m=+4910.319354602" watchObservedRunningTime="2025-11-23 21:30:02.177038113 +0000 UTC m=+4910.326079069" Nov 23 21:30:02 crc kubenswrapper[4726]: I1123 21:30:02.274652 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mvtwf_80b8ce4c-915f-424b-99c4-a16f0d9a1fc8/ssh-known-hosts-edpm-deployment/0.log" Nov 23 21:30:02 crc kubenswrapper[4726]: I1123 21:30:02.318980 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_efb73145-b382-4a15-89fe-6b74bc9d3594/tempest-tests-tempest-tests-runner/0.log" Nov 23 21:30:02 crc kubenswrapper[4726]: I1123 21:30:02.626986 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_eeaefcf2-c7a8-4b74-b883-9b05a5e01a23/test-operator-logs-container/0.log" Nov 23 21:30:02 crc kubenswrapper[4726]: I1123 21:30:02.729561 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ffnnn_da03c2c2-f43a-4f51-9c87-8e2f6b0af367/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 23 21:30:03 crc kubenswrapper[4726]: I1123 21:30:03.163957 4726 generic.go:334] "Generic (PLEG): container finished" podID="f07a9332-f2e2-41c7-8707-ac62bd160114" containerID="0fc593c5e3c575671e09a72cdb4654986643642d745080b8502c0b5cfa91c8f1" exitCode=0 Nov 23 21:30:03 crc kubenswrapper[4726]: I1123 21:30:03.163999 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" event={"ID":"f07a9332-f2e2-41c7-8707-ac62bd160114","Type":"ContainerDied","Data":"0fc593c5e3c575671e09a72cdb4654986643642d745080b8502c0b5cfa91c8f1"} Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.613412 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.778142 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft6pr\" (UniqueName: \"kubernetes.io/projected/f07a9332-f2e2-41c7-8707-ac62bd160114-kube-api-access-ft6pr\") pod \"f07a9332-f2e2-41c7-8707-ac62bd160114\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.778711 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f07a9332-f2e2-41c7-8707-ac62bd160114-secret-volume\") pod \"f07a9332-f2e2-41c7-8707-ac62bd160114\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.779038 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f07a9332-f2e2-41c7-8707-ac62bd160114-config-volume\") pod \"f07a9332-f2e2-41c7-8707-ac62bd160114\" (UID: \"f07a9332-f2e2-41c7-8707-ac62bd160114\") " Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.781113 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f07a9332-f2e2-41c7-8707-ac62bd160114-config-volume" (OuterVolumeSpecName: "config-volume") pod "f07a9332-f2e2-41c7-8707-ac62bd160114" (UID: "f07a9332-f2e2-41c7-8707-ac62bd160114"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.787854 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07a9332-f2e2-41c7-8707-ac62bd160114-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f07a9332-f2e2-41c7-8707-ac62bd160114" (UID: "f07a9332-f2e2-41c7-8707-ac62bd160114"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.812170 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f07a9332-f2e2-41c7-8707-ac62bd160114-kube-api-access-ft6pr" (OuterVolumeSpecName: "kube-api-access-ft6pr") pod "f07a9332-f2e2-41c7-8707-ac62bd160114" (UID: "f07a9332-f2e2-41c7-8707-ac62bd160114"). InnerVolumeSpecName "kube-api-access-ft6pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.881323 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f07a9332-f2e2-41c7-8707-ac62bd160114-config-volume\") on node \"crc\" DevicePath \"\"" Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.881357 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft6pr\" (UniqueName: \"kubernetes.io/projected/f07a9332-f2e2-41c7-8707-ac62bd160114-kube-api-access-ft6pr\") on node \"crc\" DevicePath \"\"" Nov 23 21:30:04 crc kubenswrapper[4726]: I1123 21:30:04.881368 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f07a9332-f2e2-41c7-8707-ac62bd160114-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 23 21:30:05 crc kubenswrapper[4726]: I1123 21:30:05.227113 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" event={"ID":"f07a9332-f2e2-41c7-8707-ac62bd160114","Type":"ContainerDied","Data":"e43feee03109e5627964ce8a8ae4a21a863f38f11b9bb14093936485a7ecfdb5"} Nov 23 21:30:05 crc kubenswrapper[4726]: I1123 21:30:05.227188 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e43feee03109e5627964ce8a8ae4a21a863f38f11b9bb14093936485a7ecfdb5" Nov 23 21:30:05 crc kubenswrapper[4726]: I1123 21:30:05.227223 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29398890-8pjcw" Nov 23 21:30:05 crc kubenswrapper[4726]: I1123 21:30:05.695239 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw"] Nov 23 21:30:05 crc kubenswrapper[4726]: I1123 21:30:05.701099 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29398845-cjtfw"] Nov 23 21:30:06 crc kubenswrapper[4726]: I1123 21:30:06.612902 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31" path="/var/lib/kubelet/pods/10ad4a6d-d9fa-47b9-8b5d-1c5df7269a31/volumes" Nov 23 21:30:09 crc kubenswrapper[4726]: I1123 21:30:09.078074 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:30:09 crc kubenswrapper[4726]: I1123 21:30:09.078571 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:30:18 crc kubenswrapper[4726]: I1123 21:30:18.250337 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_319fe004-3d9c-4bb9-bb8e-114c8949bcc7/memcached/0.log" Nov 23 21:30:21 crc kubenswrapper[4726]: I1123 21:30:21.996236 4726 scope.go:117] "RemoveContainer" containerID="2f21f64c85620c4c9f4799f6a51fc2c54515e2ff652a4f3053b1a57694e663bd" Nov 23 21:30:35 crc kubenswrapper[4726]: I1123 21:30:35.277199 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/util/0.log" Nov 23 21:30:35 crc kubenswrapper[4726]: I1123 21:30:35.443298 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/pull/0.log" Nov 23 21:30:35 crc kubenswrapper[4726]: I1123 21:30:35.474556 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/util/0.log" Nov 23 21:30:35 crc kubenswrapper[4726]: I1123 21:30:35.486435 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/pull/0.log" Nov 23 21:30:35 crc kubenswrapper[4726]: I1123 21:30:35.735269 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/util/0.log" Nov 23 21:30:35 crc kubenswrapper[4726]: I1123 21:30:35.761141 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/extract/0.log" Nov 23 21:30:35 crc kubenswrapper[4726]: I1123 21:30:35.805987 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287c57gr_6a77b1d9-7f6f-4ea1-847a-393ba021319f/pull/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.004617 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-q7qxp_d481bfb6-f28c-49ed-8efe-7fc35dc6d608/kube-rbac-proxy/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.012937 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-xt9t6_a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7/kube-rbac-proxy/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.031300 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-q7qxp_d481bfb6-f28c-49ed-8efe-7fc35dc6d608/manager/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.191733 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-xt9t6_a8dd7de4-95d6-4e3e-855c-3be0c1ed1bc7/manager/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.245900 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-wh58z_2e4ac3d0-65ba-4140-9a03-c269dfebad67/kube-rbac-proxy/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.281036 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-wh58z_2e4ac3d0-65ba-4140-9a03-c269dfebad67/manager/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.473665 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-6d8lj_bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb/kube-rbac-proxy/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.552237 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-6d8lj_bf353aea-5f6e-4aa1-9a83-ac4e9b3edbfb/manager/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.728175 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-jc55l_9c05c149-b530-4ca3-b201-309826c5e095/kube-rbac-proxy/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.744008 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-jc55l_9c05c149-b530-4ca3-b201-309826c5e095/manager/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.799047 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-574zb_c3279a0f-e448-47b2-91b3-6de329d77cec/kube-rbac-proxy/0.log" Nov 23 21:30:36 crc kubenswrapper[4726]: I1123 21:30:36.950661 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-574zb_c3279a0f-e448-47b2-91b3-6de329d77cec/manager/0.log" Nov 23 21:30:37 crc kubenswrapper[4726]: I1123 21:30:37.057472 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-w4lxk_765e2946-d293-4840-baf2-96a478748a25/kube-rbac-proxy/0.log" Nov 23 21:30:37 crc kubenswrapper[4726]: I1123 21:30:37.117050 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-w4lxk_765e2946-d293-4840-baf2-96a478748a25/manager/0.log" Nov 23 21:30:37 crc kubenswrapper[4726]: I1123 21:30:37.158136 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-m4xc4_c53c8659-c283-4c78-88b4-b60f8b040dd7/kube-rbac-proxy/0.log" Nov 23 21:30:37 crc kubenswrapper[4726]: I1123 21:30:37.394706 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-m4xc4_c53c8659-c283-4c78-88b4-b60f8b040dd7/manager/0.log" Nov 23 21:30:37 crc kubenswrapper[4726]: I1123 21:30:37.595402 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-s7746_76e7ee34-d8e7-4d5b-b79e-acd11b2b0735/kube-rbac-proxy/0.log" Nov 23 21:30:37 crc kubenswrapper[4726]: I1123 21:30:37.713319 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-s7746_76e7ee34-d8e7-4d5b-b79e-acd11b2b0735/manager/0.log" Nov 23 21:30:37 crc kubenswrapper[4726]: I1123 21:30:37.918613 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-5s9v2_a653d1f5-fe23-4a55-9422-33964eeb8488/kube-rbac-proxy/0.log" Nov 23 21:30:38 crc kubenswrapper[4726]: I1123 21:30:38.022006 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-5s9v2_a653d1f5-fe23-4a55-9422-33964eeb8488/manager/0.log" Nov 23 21:30:38 crc kubenswrapper[4726]: I1123 21:30:38.331562 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-8dz5d_b89cb12d-b390-40ed-908f-e1095fe0c62b/kube-rbac-proxy/0.log" Nov 23 21:30:38 crc kubenswrapper[4726]: I1123 21:30:38.353974 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-8dz5d_b89cb12d-b390-40ed-908f-e1095fe0c62b/manager/0.log" Nov 23 21:30:38 crc kubenswrapper[4726]: I1123 21:30:38.469040 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-4p7kg_4301d05b-4769-4c53-b557-bd6aa37efa68/kube-rbac-proxy/0.log" Nov 23 21:30:38 crc kubenswrapper[4726]: I1123 21:30:38.559918 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-4p7kg_4301d05b-4769-4c53-b557-bd6aa37efa68/manager/0.log" Nov 23 21:30:38 crc kubenswrapper[4726]: I1123 21:30:38.680928 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-qvx6m_384a1b22-97f2-4f0b-ad98-52adbbf5178b/kube-rbac-proxy/0.log" Nov 23 21:30:38 crc kubenswrapper[4726]: I1123 21:30:38.847602 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-qvx6m_384a1b22-97f2-4f0b-ad98-52adbbf5178b/manager/0.log" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.004182 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-7lwwg_20b0ca2b-a3ee-45c9-8b34-b476245a4ed6/kube-rbac-proxy/0.log" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.045178 4726 patch_prober.go:28] interesting pod/machine-config-daemon-c58qk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.045232 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.045280 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.046081 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d"} pod="openshift-machine-config-operator/machine-config-daemon-c58qk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.046135 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerName="machine-config-daemon" containerID="cri-o://cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" gracePeriod=600 Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.080505 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-7lwwg_20b0ca2b-a3ee-45c9-8b34-b476245a4ed6/manager/0.log" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.146234 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44hnx49_ce7ed28d-d56b-44de-b207-7c47ab684921/kube-rbac-proxy/0.log" Nov 23 21:30:39 crc kubenswrapper[4726]: E1123 21:30:39.178453 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.351016 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44hnx49_ce7ed28d-d56b-44de-b207-7c47ab684921/manager/0.log" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.456023 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-t9ptk_706d33da-dac2-4403-a4d8-d8e7f04f772b/kube-rbac-proxy/0.log" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.537152 4726 generic.go:334] "Generic (PLEG): container finished" podID="2e3ac186-9f76-4774-8e04-fb00add1eb72" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" exitCode=0 Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.537193 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerDied","Data":"cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d"} Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.537226 4726 scope.go:117] "RemoveContainer" containerID="d78047930e4aba8b59f1ecbdc499155597ea2347de6f9220fe2147928a3ef4b3" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.538022 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:30:39 crc kubenswrapper[4726]: E1123 21:30:39.538358 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:30:39 crc kubenswrapper[4726]: I1123 21:30:39.809352 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-v5bmf_2a88699d-09b3-458d-bbbd-03834f3c6978/kube-rbac-proxy/0.log" Nov 23 21:30:40 crc kubenswrapper[4726]: I1123 21:30:40.120155 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-v5bmf_2a88699d-09b3-458d-bbbd-03834f3c6978/operator/0.log" Nov 23 21:30:40 crc kubenswrapper[4726]: I1123 21:30:40.224529 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-whxfc_224802b7-2607-4e77-95e0-d7e02023c20b/registry-server/0.log" Nov 23 21:30:40 crc kubenswrapper[4726]: I1123 21:30:40.357923 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-2k5zx_6e692743-eb35-4a91-97af-8d86cf694470/kube-rbac-proxy/0.log" Nov 23 21:30:40 crc kubenswrapper[4726]: I1123 21:30:40.575098 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-t9ptk_706d33da-dac2-4403-a4d8-d8e7f04f772b/manager/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.056685 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-2k5zx_6e692743-eb35-4a91-97af-8d86cf694470/manager/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.068152 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-p7r99_42d03b10-01f7-4963-a97b-f3d443a1cdf7/manager/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.156647 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-p7r99_42d03b10-01f7-4963-a97b-f3d443a1cdf7/kube-rbac-proxy/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.431001 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-frss6_cfdab430-7dd4-4612-a65b-b381b5b58115/kube-rbac-proxy/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.469899 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-kvvhv_e191e3fa-cd47-4c3d-b4d4-93a24db6ffc9/operator/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.546837 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-frss6_cfdab430-7dd4-4612-a65b-b381b5b58115/manager/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.780765 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-527bk_74d488f0-9a23-4774-94e7-75605573ce43/kube-rbac-proxy/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.842932 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-527bk_74d488f0-9a23-4774-94e7-75605573ce43/manager/0.log" Nov 23 21:30:41 crc kubenswrapper[4726]: I1123 21:30:41.844846 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-5w2nt_c22e9617-abf8-4370-ba10-31e967257270/kube-rbac-proxy/0.log" Nov 23 21:30:42 crc kubenswrapper[4726]: I1123 21:30:42.011584 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-5w2nt_c22e9617-abf8-4370-ba10-31e967257270/manager/0.log" Nov 23 21:30:42 crc kubenswrapper[4726]: I1123 21:30:42.385851 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-7x7b7_023e8819-c436-4592-a6f6-ea3448733363/kube-rbac-proxy/0.log" Nov 23 21:30:42 crc kubenswrapper[4726]: I1123 21:30:42.433464 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-7x7b7_023e8819-c436-4592-a6f6-ea3448733363/manager/0.log" Nov 23 21:30:50 crc kubenswrapper[4726]: I1123 21:30:50.590044 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:30:50 crc kubenswrapper[4726]: E1123 21:30:50.590952 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:31:03 crc kubenswrapper[4726]: I1123 21:31:03.317557 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-hlvv6_4ff399e8-7c2b-497c-aec8-96383b2f92fb/control-plane-machine-set-operator/0.log" Nov 23 21:31:03 crc kubenswrapper[4726]: I1123 21:31:03.508739 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vfgwr_002cee03-8433-4e0d-b648-45300fadd459/machine-api-operator/0.log" Nov 23 21:31:03 crc kubenswrapper[4726]: I1123 21:31:03.518688 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vfgwr_002cee03-8433-4e0d-b648-45300fadd459/kube-rbac-proxy/0.log" Nov 23 21:31:03 crc kubenswrapper[4726]: I1123 21:31:03.589387 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:31:03 crc kubenswrapper[4726]: E1123 21:31:03.589981 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:31:15 crc kubenswrapper[4726]: I1123 21:31:15.999138 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vp4hq_3c7ee4c2-02fa-4cf6-8ad6-580e7ca041bd/cert-manager-controller/0.log" Nov 23 21:31:16 crc kubenswrapper[4726]: I1123 21:31:16.150619 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-dcnpn_b67043f1-eb7f-4239-b3a2-1f328e5f681f/cert-manager-cainjector/0.log" Nov 23 21:31:16 crc kubenswrapper[4726]: I1123 21:31:16.174120 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-llcjm_d7287208-3876-4ffe-b893-a588d7383fe8/cert-manager-webhook/0.log" Nov 23 21:31:18 crc kubenswrapper[4726]: I1123 21:31:18.589709 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:31:18 crc kubenswrapper[4726]: E1123 21:31:18.590493 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:31:29 crc kubenswrapper[4726]: I1123 21:31:29.590664 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:31:29 crc kubenswrapper[4726]: E1123 21:31:29.591773 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:31:30 crc kubenswrapper[4726]: I1123 21:31:30.553204 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-98ljz_6dc6c946-84ef-4ba1-8753-efd769c877c0/nmstate-console-plugin/0.log" Nov 23 21:31:30 crc kubenswrapper[4726]: I1123 21:31:30.655520 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-w9sld_97c85373-c730-40fd-8b73-0b163c59e6ba/nmstate-handler/0.log" Nov 23 21:31:30 crc kubenswrapper[4726]: I1123 21:31:30.759946 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-pnnn6_fdc047b2-c091-49cc-8a3c-9663b20e7cd1/kube-rbac-proxy/0.log" Nov 23 21:31:30 crc kubenswrapper[4726]: I1123 21:31:30.787638 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-pnnn6_fdc047b2-c091-49cc-8a3c-9663b20e7cd1/nmstate-metrics/0.log" Nov 23 21:31:30 crc kubenswrapper[4726]: I1123 21:31:30.946819 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-w8pws_0debdf86-4fa6-46aa-8ffc-513dbb889153/nmstate-operator/0.log" Nov 23 21:31:31 crc kubenswrapper[4726]: I1123 21:31:31.003454 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-jflhl_f89596fa-328e-483b-a5dd-34a241972f63/nmstate-webhook/0.log" Nov 23 21:31:40 crc kubenswrapper[4726]: I1123 21:31:40.591154 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:31:40 crc kubenswrapper[4726]: E1123 21:31:40.592189 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:31:46 crc kubenswrapper[4726]: I1123 21:31:46.423713 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-2vb27_dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0/kube-rbac-proxy/0.log" Nov 23 21:31:46 crc kubenswrapper[4726]: I1123 21:31:46.439920 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-2vb27_dad5ab40-57d4-470f-ba2e-0f1bb21ae9e0/controller/0.log" Nov 23 21:31:46 crc kubenswrapper[4726]: I1123 21:31:46.932965 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-frr-files/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.095265 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-reloader/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.153550 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-frr-files/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.188374 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-metrics/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.194690 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-reloader/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.367525 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-metrics/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.417193 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-frr-files/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.427125 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-reloader/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.460793 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-metrics/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.629599 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-reloader/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.668179 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-frr-files/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.729105 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/controller/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.744412 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/cp-metrics/0.log" Nov 23 21:31:47 crc kubenswrapper[4726]: I1123 21:31:47.934659 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/frr-metrics/0.log" Nov 23 21:31:48 crc kubenswrapper[4726]: I1123 21:31:48.003789 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/kube-rbac-proxy/0.log" Nov 23 21:31:48 crc kubenswrapper[4726]: I1123 21:31:48.056992 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/kube-rbac-proxy-frr/0.log" Nov 23 21:31:48 crc kubenswrapper[4726]: I1123 21:31:48.179864 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/reloader/0.log" Nov 23 21:31:48 crc kubenswrapper[4726]: I1123 21:31:48.889413 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-mlnqp_3130c09b-7002-4448-bdd4-87b73f70ff99/frr-k8s-webhook-server/0.log" Nov 23 21:31:49 crc kubenswrapper[4726]: I1123 21:31:49.117003 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-64f7f896bc-tb44t_be91600b-06c3-4acc-a0d6-d7e732a00228/manager/0.log" Nov 23 21:31:49 crc kubenswrapper[4726]: I1123 21:31:49.194562 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c66969f79-8m676_bf3b7cb9-1dc8-4966-b15c-71d77aa2751b/webhook-server/0.log" Nov 23 21:31:49 crc kubenswrapper[4726]: I1123 21:31:49.254955 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-b8hdl_87083cfa-76dd-42ca-a8e2-a3900f4f6bd5/frr/0.log" Nov 23 21:31:49 crc kubenswrapper[4726]: I1123 21:31:49.455505 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zkvn5_d99ab20e-6022-4d7e-8ff2-571cd0961a08/kube-rbac-proxy/0.log" Nov 23 21:31:49 crc kubenswrapper[4726]: I1123 21:31:49.651189 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zkvn5_d99ab20e-6022-4d7e-8ff2-571cd0961a08/speaker/0.log" Nov 23 21:31:51 crc kubenswrapper[4726]: I1123 21:31:51.589565 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:31:51 crc kubenswrapper[4726]: E1123 21:31:51.590222 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:32:02 crc kubenswrapper[4726]: I1123 21:32:02.596410 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:32:02 crc kubenswrapper[4726]: E1123 21:32:02.598239 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:32:02 crc kubenswrapper[4726]: I1123 21:32:02.690952 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/util/0.log" Nov 23 21:32:02 crc kubenswrapper[4726]: I1123 21:32:02.876740 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/pull/0.log" Nov 23 21:32:02 crc kubenswrapper[4726]: I1123 21:32:02.907480 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/util/0.log" Nov 23 21:32:02 crc kubenswrapper[4726]: I1123 21:32:02.928555 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/pull/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.097338 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/util/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.106179 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/pull/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.167879 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ez928h_71acdd00-d9b5-4b6a-a454-f8f65c4a5365/extract/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.288152 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-utilities/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.505505 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-content/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.507381 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-utilities/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.555574 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-content/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.762952 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-content/0.log" Nov 23 21:32:03 crc kubenswrapper[4726]: I1123 21:32:03.864714 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/extract-utilities/0.log" Nov 23 21:32:04 crc kubenswrapper[4726]: I1123 21:32:04.063928 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-utilities/0.log" Nov 23 21:32:04 crc kubenswrapper[4726]: I1123 21:32:04.293750 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-content/0.log" Nov 23 21:32:04 crc kubenswrapper[4726]: I1123 21:32:04.302127 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-utilities/0.log" Nov 23 21:32:04 crc kubenswrapper[4726]: I1123 21:32:04.388402 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8qfs7_5bb09cfa-b62c-4b90-9cbc-2309be53c15a/registry-server/0.log" Nov 23 21:32:04 crc kubenswrapper[4726]: I1123 21:32:04.434994 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-content/0.log" Nov 23 21:32:04 crc kubenswrapper[4726]: I1123 21:32:04.579138 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-content/0.log" Nov 23 21:32:04 crc kubenswrapper[4726]: I1123 21:32:04.580378 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/extract-utilities/0.log" Nov 23 21:32:04 crc kubenswrapper[4726]: I1123 21:32:04.889547 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/util/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.085477 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/pull/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.211920 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/util/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.255805 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/pull/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.340464 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-r7nwg_7459f94c-2ed1-470a-a308-1377af995ab7/registry-server/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.517299 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/pull/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.576328 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/util/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.600797 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c67qvrx_ea8a3056-cbc8-4a11-9144-7f6a939e697d/extract/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.714067 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7rvp2_cfd53c53-6459-4a5d-b8e4-bf47ebb6da34/marketplace-operator/0.log" Nov 23 21:32:05 crc kubenswrapper[4726]: I1123 21:32:05.789775 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-utilities/0.log" Nov 23 21:32:06 crc kubenswrapper[4726]: I1123 21:32:06.085806 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-content/0.log" Nov 23 21:32:06 crc kubenswrapper[4726]: I1123 21:32:06.087182 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-content/0.log" Nov 23 21:32:06 crc kubenswrapper[4726]: I1123 21:32:06.093314 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-utilities/0.log" Nov 23 21:32:06 crc kubenswrapper[4726]: I1123 21:32:06.896052 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-utilities/0.log" Nov 23 21:32:06 crc kubenswrapper[4726]: I1123 21:32:06.941324 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/extract-content/0.log" Nov 23 21:32:06 crc kubenswrapper[4726]: I1123 21:32:06.971493 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-utilities/0.log" Nov 23 21:32:07 crc kubenswrapper[4726]: I1123 21:32:07.107960 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b28ts_75ecd57d-a57d-44fc-811a-1d30d925e93e/registry-server/0.log" Nov 23 21:32:07 crc kubenswrapper[4726]: I1123 21:32:07.246564 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-content/0.log" Nov 23 21:32:07 crc kubenswrapper[4726]: I1123 21:32:07.278442 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-content/0.log" Nov 23 21:32:07 crc kubenswrapper[4726]: I1123 21:32:07.316082 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-utilities/0.log" Nov 23 21:32:07 crc kubenswrapper[4726]: I1123 21:32:07.466807 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-content/0.log" Nov 23 21:32:07 crc kubenswrapper[4726]: I1123 21:32:07.482050 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/extract-utilities/0.log" Nov 23 21:32:08 crc kubenswrapper[4726]: I1123 21:32:08.142360 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t8776_d5073784-d975-4fce-805f-7216d9356df8/registry-server/0.log" Nov 23 21:32:16 crc kubenswrapper[4726]: I1123 21:32:16.592220 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:32:16 crc kubenswrapper[4726]: E1123 21:32:16.592917 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:32:31 crc kubenswrapper[4726]: I1123 21:32:31.588794 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:32:31 crc kubenswrapper[4726]: E1123 21:32:31.589487 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:32:42 crc kubenswrapper[4726]: I1123 21:32:42.599834 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:32:42 crc kubenswrapper[4726]: E1123 21:32:42.606933 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:32:57 crc kubenswrapper[4726]: I1123 21:32:57.589519 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:32:57 crc kubenswrapper[4726]: E1123 21:32:57.590249 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:33:11 crc kubenswrapper[4726]: I1123 21:33:11.590224 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:33:11 crc kubenswrapper[4726]: E1123 21:33:11.591141 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:33:23 crc kubenswrapper[4726]: I1123 21:33:23.590606 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:33:23 crc kubenswrapper[4726]: E1123 21:33:23.592132 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:33:34 crc kubenswrapper[4726]: I1123 21:33:34.589852 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:33:34 crc kubenswrapper[4726]: E1123 21:33:34.590714 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:33:49 crc kubenswrapper[4726]: I1123 21:33:49.588860 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:33:49 crc kubenswrapper[4726]: E1123 21:33:49.591014 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:34:01 crc kubenswrapper[4726]: I1123 21:34:01.588721 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:34:01 crc kubenswrapper[4726]: E1123 21:34:01.590159 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:34:13 crc kubenswrapper[4726]: I1123 21:34:13.589660 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:34:13 crc kubenswrapper[4726]: E1123 21:34:13.591679 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.761004 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8vplq"] Nov 23 21:34:17 crc kubenswrapper[4726]: E1123 21:34:17.762704 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f07a9332-f2e2-41c7-8707-ac62bd160114" containerName="collect-profiles" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.762724 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f07a9332-f2e2-41c7-8707-ac62bd160114" containerName="collect-profiles" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.763150 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f07a9332-f2e2-41c7-8707-ac62bd160114" containerName="collect-profiles" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.766135 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.788474 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vplq"] Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.847739 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-catalog-content\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.847801 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-utilities\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.847966 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxzt6\" (UniqueName: \"kubernetes.io/projected/ea243882-404a-4dea-8029-5cf2d1ab0dfd-kube-api-access-kxzt6\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.950839 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-catalog-content\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.950933 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-utilities\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.951099 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxzt6\" (UniqueName: \"kubernetes.io/projected/ea243882-404a-4dea-8029-5cf2d1ab0dfd-kube-api-access-kxzt6\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.957284 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-utilities\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.960304 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-catalog-content\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:17 crc kubenswrapper[4726]: I1123 21:34:17.986071 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxzt6\" (UniqueName: \"kubernetes.io/projected/ea243882-404a-4dea-8029-5cf2d1ab0dfd-kube-api-access-kxzt6\") pod \"certified-operators-8vplq\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:18 crc kubenswrapper[4726]: I1123 21:34:18.279694 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:18 crc kubenswrapper[4726]: I1123 21:34:18.965149 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vplq"] Nov 23 21:34:19 crc kubenswrapper[4726]: I1123 21:34:19.439474 4726 generic.go:334] "Generic (PLEG): container finished" podID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerID="4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5" exitCode=0 Nov 23 21:34:19 crc kubenswrapper[4726]: I1123 21:34:19.439602 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vplq" event={"ID":"ea243882-404a-4dea-8029-5cf2d1ab0dfd","Type":"ContainerDied","Data":"4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5"} Nov 23 21:34:19 crc kubenswrapper[4726]: I1123 21:34:19.439780 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vplq" event={"ID":"ea243882-404a-4dea-8029-5cf2d1ab0dfd","Type":"ContainerStarted","Data":"d044bd2fe40230621ee75ea36d3e8ce844caadc17ff664aeb2ca5d526bccf919"} Nov 23 21:34:19 crc kubenswrapper[4726]: I1123 21:34:19.442257 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 23 21:34:20 crc kubenswrapper[4726]: I1123 21:34:20.448815 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vplq" event={"ID":"ea243882-404a-4dea-8029-5cf2d1ab0dfd","Type":"ContainerStarted","Data":"136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c"} Nov 23 21:34:22 crc kubenswrapper[4726]: I1123 21:34:22.467586 4726 generic.go:334] "Generic (PLEG): container finished" podID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerID="136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c" exitCode=0 Nov 23 21:34:22 crc kubenswrapper[4726]: I1123 21:34:22.468108 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vplq" event={"ID":"ea243882-404a-4dea-8029-5cf2d1ab0dfd","Type":"ContainerDied","Data":"136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c"} Nov 23 21:34:23 crc kubenswrapper[4726]: I1123 21:34:23.485826 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vplq" event={"ID":"ea243882-404a-4dea-8029-5cf2d1ab0dfd","Type":"ContainerStarted","Data":"a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83"} Nov 23 21:34:23 crc kubenswrapper[4726]: I1123 21:34:23.513720 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8vplq" podStartSLOduration=3.088234697 podStartE2EDuration="6.513698993s" podCreationTimestamp="2025-11-23 21:34:17 +0000 UTC" firstStartedPulling="2025-11-23 21:34:19.44206321 +0000 UTC m=+5167.591104166" lastFinishedPulling="2025-11-23 21:34:22.867527506 +0000 UTC m=+5171.016568462" observedRunningTime="2025-11-23 21:34:23.505334059 +0000 UTC m=+5171.654375035" watchObservedRunningTime="2025-11-23 21:34:23.513698993 +0000 UTC m=+5171.662739959" Nov 23 21:34:24 crc kubenswrapper[4726]: I1123 21:34:24.588947 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:34:24 crc kubenswrapper[4726]: E1123 21:34:24.591212 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:34:27 crc kubenswrapper[4726]: I1123 21:34:27.521376 4726 generic.go:334] "Generic (PLEG): container finished" podID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerID="cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63" exitCode=0 Nov 23 21:34:27 crc kubenswrapper[4726]: I1123 21:34:27.521490 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" event={"ID":"637adc9a-b276-423b-a8ed-bf6f3143ec97","Type":"ContainerDied","Data":"cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63"} Nov 23 21:34:27 crc kubenswrapper[4726]: I1123 21:34:27.522378 4726 scope.go:117] "RemoveContainer" containerID="cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63" Nov 23 21:34:28 crc kubenswrapper[4726]: I1123 21:34:28.184494 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rtcdk_must-gather-xkxgs_637adc9a-b276-423b-a8ed-bf6f3143ec97/gather/0.log" Nov 23 21:34:28 crc kubenswrapper[4726]: I1123 21:34:28.280161 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:28 crc kubenswrapper[4726]: I1123 21:34:28.281581 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:28 crc kubenswrapper[4726]: I1123 21:34:28.344589 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:28 crc kubenswrapper[4726]: I1123 21:34:28.584351 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:28 crc kubenswrapper[4726]: I1123 21:34:28.645046 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vplq"] Nov 23 21:34:30 crc kubenswrapper[4726]: I1123 21:34:30.604515 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8vplq" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerName="registry-server" containerID="cri-o://a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83" gracePeriod=2 Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.123228 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.143521 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxzt6\" (UniqueName: \"kubernetes.io/projected/ea243882-404a-4dea-8029-5cf2d1ab0dfd-kube-api-access-kxzt6\") pod \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.143696 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-utilities\") pod \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.143943 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-catalog-content\") pod \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\" (UID: \"ea243882-404a-4dea-8029-5cf2d1ab0dfd\") " Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.144740 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-utilities" (OuterVolumeSpecName: "utilities") pod "ea243882-404a-4dea-8029-5cf2d1ab0dfd" (UID: "ea243882-404a-4dea-8029-5cf2d1ab0dfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.182329 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea243882-404a-4dea-8029-5cf2d1ab0dfd-kube-api-access-kxzt6" (OuterVolumeSpecName: "kube-api-access-kxzt6") pod "ea243882-404a-4dea-8029-5cf2d1ab0dfd" (UID: "ea243882-404a-4dea-8029-5cf2d1ab0dfd"). InnerVolumeSpecName "kube-api-access-kxzt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.210795 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea243882-404a-4dea-8029-5cf2d1ab0dfd" (UID: "ea243882-404a-4dea-8029-5cf2d1ab0dfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.245348 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.245382 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxzt6\" (UniqueName: \"kubernetes.io/projected/ea243882-404a-4dea-8029-5cf2d1ab0dfd-kube-api-access-kxzt6\") on node \"crc\" DevicePath \"\"" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.245409 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea243882-404a-4dea-8029-5cf2d1ab0dfd-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.614101 4726 generic.go:334] "Generic (PLEG): container finished" podID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerID="a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83" exitCode=0 Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.614138 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vplq" event={"ID":"ea243882-404a-4dea-8029-5cf2d1ab0dfd","Type":"ContainerDied","Data":"a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83"} Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.614162 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vplq" event={"ID":"ea243882-404a-4dea-8029-5cf2d1ab0dfd","Type":"ContainerDied","Data":"d044bd2fe40230621ee75ea36d3e8ce844caadc17ff664aeb2ca5d526bccf919"} Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.614178 4726 scope.go:117] "RemoveContainer" containerID="a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.614300 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vplq" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.641585 4726 scope.go:117] "RemoveContainer" containerID="136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.650217 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vplq"] Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.662715 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8vplq"] Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.692548 4726 scope.go:117] "RemoveContainer" containerID="4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.733622 4726 scope.go:117] "RemoveContainer" containerID="a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83" Nov 23 21:34:31 crc kubenswrapper[4726]: E1123 21:34:31.734121 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83\": container with ID starting with a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83 not found: ID does not exist" containerID="a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.734150 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83"} err="failed to get container status \"a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83\": rpc error: code = NotFound desc = could not find container \"a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83\": container with ID starting with a330a5db76409fb7780761073fb96e105785e18032bd6a7dce459ac076d64b83 not found: ID does not exist" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.734171 4726 scope.go:117] "RemoveContainer" containerID="136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c" Nov 23 21:34:31 crc kubenswrapper[4726]: E1123 21:34:31.734509 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c\": container with ID starting with 136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c not found: ID does not exist" containerID="136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.734530 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c"} err="failed to get container status \"136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c\": rpc error: code = NotFound desc = could not find container \"136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c\": container with ID starting with 136c349b70ba76eb256cbd20af0faf58459585984dfbd0baf5d822b7dd0b7e3c not found: ID does not exist" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.734542 4726 scope.go:117] "RemoveContainer" containerID="4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5" Nov 23 21:34:31 crc kubenswrapper[4726]: E1123 21:34:31.734833 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5\": container with ID starting with 4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5 not found: ID does not exist" containerID="4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5" Nov 23 21:34:31 crc kubenswrapper[4726]: I1123 21:34:31.734854 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5"} err="failed to get container status \"4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5\": rpc error: code = NotFound desc = could not find container \"4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5\": container with ID starting with 4f49cf3009f7460d7337d39d0055560b98e4b1fe79d26c8594ea01e081d5eef5 not found: ID does not exist" Nov 23 21:34:32 crc kubenswrapper[4726]: I1123 21:34:32.601093 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" path="/var/lib/kubelet/pods/ea243882-404a-4dea-8029-5cf2d1ab0dfd/volumes" Nov 23 21:34:37 crc kubenswrapper[4726]: I1123 21:34:37.589397 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:34:37 crc kubenswrapper[4726]: E1123 21:34:37.591482 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.189358 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rtcdk/must-gather-xkxgs"] Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.190138 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" podUID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerName="copy" containerID="cri-o://f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21" gracePeriod=2 Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.207093 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rtcdk/must-gather-xkxgs"] Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.653113 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rtcdk_must-gather-xkxgs_637adc9a-b276-423b-a8ed-bf6f3143ec97/copy/0.log" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.653771 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.719279 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gthfx\" (UniqueName: \"kubernetes.io/projected/637adc9a-b276-423b-a8ed-bf6f3143ec97-kube-api-access-gthfx\") pod \"637adc9a-b276-423b-a8ed-bf6f3143ec97\" (UID: \"637adc9a-b276-423b-a8ed-bf6f3143ec97\") " Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.727099 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/637adc9a-b276-423b-a8ed-bf6f3143ec97-kube-api-access-gthfx" (OuterVolumeSpecName: "kube-api-access-gthfx") pod "637adc9a-b276-423b-a8ed-bf6f3143ec97" (UID: "637adc9a-b276-423b-a8ed-bf6f3143ec97"). InnerVolumeSpecName "kube-api-access-gthfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.741279 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rtcdk_must-gather-xkxgs_637adc9a-b276-423b-a8ed-bf6f3143ec97/copy/0.log" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.741641 4726 generic.go:334] "Generic (PLEG): container finished" podID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerID="f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21" exitCode=143 Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.741693 4726 scope.go:117] "RemoveContainer" containerID="f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.741797 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtcdk/must-gather-xkxgs" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.779437 4726 scope.go:117] "RemoveContainer" containerID="cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.821344 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/637adc9a-b276-423b-a8ed-bf6f3143ec97-must-gather-output\") pod \"637adc9a-b276-423b-a8ed-bf6f3143ec97\" (UID: \"637adc9a-b276-423b-a8ed-bf6f3143ec97\") " Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.821734 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gthfx\" (UniqueName: \"kubernetes.io/projected/637adc9a-b276-423b-a8ed-bf6f3143ec97-kube-api-access-gthfx\") on node \"crc\" DevicePath \"\"" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.836525 4726 scope.go:117] "RemoveContainer" containerID="f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21" Nov 23 21:34:42 crc kubenswrapper[4726]: E1123 21:34:42.836957 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21\": container with ID starting with f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21 not found: ID does not exist" containerID="f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.836988 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21"} err="failed to get container status \"f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21\": rpc error: code = NotFound desc = could not find container \"f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21\": container with ID starting with f1e9c39febfff86e162bcfc5b5a58d12f1567dfb14f613faf533a33a2e004a21 not found: ID does not exist" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.837009 4726 scope.go:117] "RemoveContainer" containerID="cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63" Nov 23 21:34:42 crc kubenswrapper[4726]: E1123 21:34:42.837285 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63\": container with ID starting with cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63 not found: ID does not exist" containerID="cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63" Nov 23 21:34:42 crc kubenswrapper[4726]: I1123 21:34:42.837307 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63"} err="failed to get container status \"cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63\": rpc error: code = NotFound desc = could not find container \"cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63\": container with ID starting with cc6afd9ff297086f18ddb75b25434ea212f9fa80c5332fffaff5a76906fb6d63 not found: ID does not exist" Nov 23 21:34:43 crc kubenswrapper[4726]: I1123 21:34:43.001000 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/637adc9a-b276-423b-a8ed-bf6f3143ec97-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "637adc9a-b276-423b-a8ed-bf6f3143ec97" (UID: "637adc9a-b276-423b-a8ed-bf6f3143ec97"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:34:43 crc kubenswrapper[4726]: I1123 21:34:43.025714 4726 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/637adc9a-b276-423b-a8ed-bf6f3143ec97-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 23 21:34:44 crc kubenswrapper[4726]: I1123 21:34:44.601637 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="637adc9a-b276-423b-a8ed-bf6f3143ec97" path="/var/lib/kubelet/pods/637adc9a-b276-423b-a8ed-bf6f3143ec97/volumes" Nov 23 21:34:49 crc kubenswrapper[4726]: I1123 21:34:49.588917 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:34:49 crc kubenswrapper[4726]: E1123 21:34:49.590892 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:35:02 crc kubenswrapper[4726]: I1123 21:35:02.595483 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:35:02 crc kubenswrapper[4726]: E1123 21:35:02.596333 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:35:16 crc kubenswrapper[4726]: I1123 21:35:16.589561 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:35:16 crc kubenswrapper[4726]: E1123 21:35:16.591630 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:35:22 crc kubenswrapper[4726]: I1123 21:35:22.123834 4726 scope.go:117] "RemoveContainer" containerID="b17aa32910d2f26e29661bbfa10031a674c532f1dff9b6a68d57145721d3367e" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.397083 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-25gqw"] Nov 23 21:35:26 crc kubenswrapper[4726]: E1123 21:35:26.397988 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerName="extract-content" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.398000 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerName="extract-content" Nov 23 21:35:26 crc kubenswrapper[4726]: E1123 21:35:26.398014 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerName="gather" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.398022 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerName="gather" Nov 23 21:35:26 crc kubenswrapper[4726]: E1123 21:35:26.398036 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerName="extract-utilities" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.398044 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerName="extract-utilities" Nov 23 21:35:26 crc kubenswrapper[4726]: E1123 21:35:26.398058 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerName="registry-server" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.398065 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerName="registry-server" Nov 23 21:35:26 crc kubenswrapper[4726]: E1123 21:35:26.398085 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerName="copy" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.398090 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerName="copy" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.398246 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerName="copy" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.398263 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="637adc9a-b276-423b-a8ed-bf6f3143ec97" containerName="gather" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.398275 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea243882-404a-4dea-8029-5cf2d1ab0dfd" containerName="registry-server" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.399545 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.409636 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cktzb\" (UniqueName: \"kubernetes.io/projected/0b864d33-4ec2-480d-aba8-37f4cde22ad4-kube-api-access-cktzb\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.409696 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-utilities\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.409751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-catalog-content\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.418793 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-25gqw"] Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.510661 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cktzb\" (UniqueName: \"kubernetes.io/projected/0b864d33-4ec2-480d-aba8-37f4cde22ad4-kube-api-access-cktzb\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.510725 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-utilities\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.510772 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-catalog-content\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.511247 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-catalog-content\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.511704 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-utilities\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.531565 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cktzb\" (UniqueName: \"kubernetes.io/projected/0b864d33-4ec2-480d-aba8-37f4cde22ad4-kube-api-access-cktzb\") pod \"redhat-operators-25gqw\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:26 crc kubenswrapper[4726]: I1123 21:35:26.723318 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:27 crc kubenswrapper[4726]: I1123 21:35:27.274558 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-25gqw"] Nov 23 21:35:28 crc kubenswrapper[4726]: I1123 21:35:28.188656 4726 generic.go:334] "Generic (PLEG): container finished" podID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerID="5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c" exitCode=0 Nov 23 21:35:28 crc kubenswrapper[4726]: I1123 21:35:28.189854 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25gqw" event={"ID":"0b864d33-4ec2-480d-aba8-37f4cde22ad4","Type":"ContainerDied","Data":"5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c"} Nov 23 21:35:28 crc kubenswrapper[4726]: I1123 21:35:28.190251 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25gqw" event={"ID":"0b864d33-4ec2-480d-aba8-37f4cde22ad4","Type":"ContainerStarted","Data":"99405e44c7a3a5e8a0b7b08b30d66822846d23478ae84b2c69461c0cfa14dea7"} Nov 23 21:35:29 crc kubenswrapper[4726]: I1123 21:35:29.203206 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25gqw" event={"ID":"0b864d33-4ec2-480d-aba8-37f4cde22ad4","Type":"ContainerStarted","Data":"d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0"} Nov 23 21:35:31 crc kubenswrapper[4726]: I1123 21:35:31.588832 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:35:31 crc kubenswrapper[4726]: E1123 21:35:31.589395 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c58qk_openshift-machine-config-operator(2e3ac186-9f76-4774-8e04-fb00add1eb72)\"" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" podUID="2e3ac186-9f76-4774-8e04-fb00add1eb72" Nov 23 21:35:33 crc kubenswrapper[4726]: I1123 21:35:33.244341 4726 generic.go:334] "Generic (PLEG): container finished" podID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerID="d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0" exitCode=0 Nov 23 21:35:33 crc kubenswrapper[4726]: I1123 21:35:33.244415 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25gqw" event={"ID":"0b864d33-4ec2-480d-aba8-37f4cde22ad4","Type":"ContainerDied","Data":"d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0"} Nov 23 21:35:34 crc kubenswrapper[4726]: I1123 21:35:34.257822 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25gqw" event={"ID":"0b864d33-4ec2-480d-aba8-37f4cde22ad4","Type":"ContainerStarted","Data":"07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178"} Nov 23 21:35:34 crc kubenswrapper[4726]: I1123 21:35:34.297486 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-25gqw" podStartSLOduration=2.740183725 podStartE2EDuration="8.297460966s" podCreationTimestamp="2025-11-23 21:35:26 +0000 UTC" firstStartedPulling="2025-11-23 21:35:28.198250802 +0000 UTC m=+5236.347291768" lastFinishedPulling="2025-11-23 21:35:33.755528063 +0000 UTC m=+5241.904569009" observedRunningTime="2025-11-23 21:35:34.27575736 +0000 UTC m=+5242.424798346" watchObservedRunningTime="2025-11-23 21:35:34.297460966 +0000 UTC m=+5242.446501932" Nov 23 21:35:36 crc kubenswrapper[4726]: I1123 21:35:36.723841 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:36 crc kubenswrapper[4726]: I1123 21:35:36.724143 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:37 crc kubenswrapper[4726]: I1123 21:35:37.768436 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-25gqw" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="registry-server" probeResult="failure" output=< Nov 23 21:35:37 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Nov 23 21:35:37 crc kubenswrapper[4726]: > Nov 23 21:35:44 crc kubenswrapper[4726]: I1123 21:35:44.589250 4726 scope.go:117] "RemoveContainer" containerID="cdf15ddebcabfdf43c6d1f6b5fe09b732b8f09867e90c084b51167b8e89fe54d" Nov 23 21:35:45 crc kubenswrapper[4726]: I1123 21:35:45.359615 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c58qk" event={"ID":"2e3ac186-9f76-4774-8e04-fb00add1eb72","Type":"ContainerStarted","Data":"d9cad364302b9b0ffb93706dec21fb017f3a3b93a36c7d9a7f143a8ade83da5d"} Nov 23 21:35:46 crc kubenswrapper[4726]: I1123 21:35:46.797078 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:46 crc kubenswrapper[4726]: I1123 21:35:46.882058 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:47 crc kubenswrapper[4726]: I1123 21:35:47.044826 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-25gqw"] Nov 23 21:35:48 crc kubenswrapper[4726]: I1123 21:35:48.389323 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-25gqw" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="registry-server" containerID="cri-o://07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178" gracePeriod=2 Nov 23 21:35:48 crc kubenswrapper[4726]: I1123 21:35:48.920475 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:48 crc kubenswrapper[4726]: I1123 21:35:48.936631 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-utilities\") pod \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " Nov 23 21:35:48 crc kubenswrapper[4726]: I1123 21:35:48.936780 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-catalog-content\") pod \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " Nov 23 21:35:48 crc kubenswrapper[4726]: I1123 21:35:48.936830 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cktzb\" (UniqueName: \"kubernetes.io/projected/0b864d33-4ec2-480d-aba8-37f4cde22ad4-kube-api-access-cktzb\") pod \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\" (UID: \"0b864d33-4ec2-480d-aba8-37f4cde22ad4\") " Nov 23 21:35:48 crc kubenswrapper[4726]: I1123 21:35:48.937472 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-utilities" (OuterVolumeSpecName: "utilities") pod "0b864d33-4ec2-480d-aba8-37f4cde22ad4" (UID: "0b864d33-4ec2-480d-aba8-37f4cde22ad4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:35:48 crc kubenswrapper[4726]: I1123 21:35:48.984191 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b864d33-4ec2-480d-aba8-37f4cde22ad4-kube-api-access-cktzb" (OuterVolumeSpecName: "kube-api-access-cktzb") pod "0b864d33-4ec2-480d-aba8-37f4cde22ad4" (UID: "0b864d33-4ec2-480d-aba8-37f4cde22ad4"). InnerVolumeSpecName "kube-api-access-cktzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.038933 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cktzb\" (UniqueName: \"kubernetes.io/projected/0b864d33-4ec2-480d-aba8-37f4cde22ad4-kube-api-access-cktzb\") on node \"crc\" DevicePath \"\"" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.038959 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.047303 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b864d33-4ec2-480d-aba8-37f4cde22ad4" (UID: "0b864d33-4ec2-480d-aba8-37f4cde22ad4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.140446 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b864d33-4ec2-480d-aba8-37f4cde22ad4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.400645 4726 generic.go:334] "Generic (PLEG): container finished" podID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerID="07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178" exitCode=0 Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.400693 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25gqw" event={"ID":"0b864d33-4ec2-480d-aba8-37f4cde22ad4","Type":"ContainerDied","Data":"07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178"} Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.400732 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25gqw" event={"ID":"0b864d33-4ec2-480d-aba8-37f4cde22ad4","Type":"ContainerDied","Data":"99405e44c7a3a5e8a0b7b08b30d66822846d23478ae84b2c69461c0cfa14dea7"} Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.400751 4726 scope.go:117] "RemoveContainer" containerID="07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.401933 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25gqw" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.449409 4726 scope.go:117] "RemoveContainer" containerID="d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.455077 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-25gqw"] Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.463023 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-25gqw"] Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.919393 4726 scope.go:117] "RemoveContainer" containerID="5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.945578 4726 scope.go:117] "RemoveContainer" containerID="07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178" Nov 23 21:35:49 crc kubenswrapper[4726]: E1123 21:35:49.946043 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178\": container with ID starting with 07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178 not found: ID does not exist" containerID="07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.946083 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178"} err="failed to get container status \"07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178\": rpc error: code = NotFound desc = could not find container \"07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178\": container with ID starting with 07115045587ccad4c7edb2869fea4ed6fe65282dfcbdf527ddac0a5050b38178 not found: ID does not exist" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.946119 4726 scope.go:117] "RemoveContainer" containerID="d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0" Nov 23 21:35:49 crc kubenswrapper[4726]: E1123 21:35:49.947194 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0\": container with ID starting with d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0 not found: ID does not exist" containerID="d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.947214 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0"} err="failed to get container status \"d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0\": rpc error: code = NotFound desc = could not find container \"d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0\": container with ID starting with d2b9c8938f46b67d916c187dc5bf4122903ace0387a71455328c3a1d58118cc0 not found: ID does not exist" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.947226 4726 scope.go:117] "RemoveContainer" containerID="5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c" Nov 23 21:35:49 crc kubenswrapper[4726]: E1123 21:35:49.947474 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c\": container with ID starting with 5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c not found: ID does not exist" containerID="5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c" Nov 23 21:35:49 crc kubenswrapper[4726]: I1123 21:35:49.947492 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c"} err="failed to get container status \"5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c\": rpc error: code = NotFound desc = could not find container \"5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c\": container with ID starting with 5a22e60af7e171ae8eed65a596cc0ae1800e3f6a5235eddf7380925060d1429c not found: ID does not exist" Nov 23 21:35:50 crc kubenswrapper[4726]: I1123 21:35:50.600774 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" path="/var/lib/kubelet/pods/0b864d33-4ec2-480d-aba8-37f4cde22ad4/volumes" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.543657 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-47n5x"] Nov 23 21:37:14 crc kubenswrapper[4726]: E1123 21:37:14.544887 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="extract-content" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.544909 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="extract-content" Nov 23 21:37:14 crc kubenswrapper[4726]: E1123 21:37:14.544954 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="extract-utilities" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.544967 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="extract-utilities" Nov 23 21:37:14 crc kubenswrapper[4726]: E1123 21:37:14.544996 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="registry-server" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.545008 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="registry-server" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.545320 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b864d33-4ec2-480d-aba8-37f4cde22ad4" containerName="registry-server" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.547679 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.557052 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47n5x"] Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.630263 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7qv5\" (UniqueName: \"kubernetes.io/projected/4411be92-5e58-429c-a1ef-cf35c7674665-kube-api-access-q7qv5\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.630821 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-catalog-content\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.630957 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-utilities\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.733661 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7qv5\" (UniqueName: \"kubernetes.io/projected/4411be92-5e58-429c-a1ef-cf35c7674665-kube-api-access-q7qv5\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.733768 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-catalog-content\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.733848 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-utilities\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.734508 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-catalog-content\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.734665 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-utilities\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.763406 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7qv5\" (UniqueName: \"kubernetes.io/projected/4411be92-5e58-429c-a1ef-cf35c7674665-kube-api-access-q7qv5\") pod \"redhat-marketplace-47n5x\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:14 crc kubenswrapper[4726]: I1123 21:37:14.880687 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:15 crc kubenswrapper[4726]: I1123 21:37:15.315989 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47n5x"] Nov 23 21:37:15 crc kubenswrapper[4726]: I1123 21:37:15.410568 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47n5x" event={"ID":"4411be92-5e58-429c-a1ef-cf35c7674665","Type":"ContainerStarted","Data":"6b9bb5194e2dc7a3fc7f0f8232130da1291b431690f47e086f604647d15904fc"} Nov 23 21:37:16 crc kubenswrapper[4726]: I1123 21:37:16.434684 4726 generic.go:334] "Generic (PLEG): container finished" podID="4411be92-5e58-429c-a1ef-cf35c7674665" containerID="f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97" exitCode=0 Nov 23 21:37:16 crc kubenswrapper[4726]: I1123 21:37:16.435142 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47n5x" event={"ID":"4411be92-5e58-429c-a1ef-cf35c7674665","Type":"ContainerDied","Data":"f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97"} Nov 23 21:37:16 crc kubenswrapper[4726]: I1123 21:37:16.945165 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cm5bq"] Nov 23 21:37:16 crc kubenswrapper[4726]: I1123 21:37:16.951669 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:16 crc kubenswrapper[4726]: I1123 21:37:16.963233 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cm5bq"] Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.081098 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-catalog-content\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.081176 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt5mq\" (UniqueName: \"kubernetes.io/projected/270843b1-4cd4-4204-a4f5-140276a7b641-kube-api-access-pt5mq\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.081253 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-utilities\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.182688 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-catalog-content\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.182751 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt5mq\" (UniqueName: \"kubernetes.io/projected/270843b1-4cd4-4204-a4f5-140276a7b641-kube-api-access-pt5mq\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.182805 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-utilities\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.183340 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-utilities\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.183341 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-catalog-content\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.212519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt5mq\" (UniqueName: \"kubernetes.io/projected/270843b1-4cd4-4204-a4f5-140276a7b641-kube-api-access-pt5mq\") pod \"community-operators-cm5bq\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.288379 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.544817 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47n5x" event={"ID":"4411be92-5e58-429c-a1ef-cf35c7674665","Type":"ContainerStarted","Data":"a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c"} Nov 23 21:37:17 crc kubenswrapper[4726]: I1123 21:37:17.900758 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cm5bq"] Nov 23 21:37:18 crc kubenswrapper[4726]: E1123 21:37:18.114951 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4411be92_5e58_429c_a1ef_cf35c7674665.slice/crio-conmon-a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c.scope\": RecentStats: unable to find data in memory cache]" Nov 23 21:37:18 crc kubenswrapper[4726]: I1123 21:37:18.554251 4726 generic.go:334] "Generic (PLEG): container finished" podID="270843b1-4cd4-4204-a4f5-140276a7b641" containerID="c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80" exitCode=0 Nov 23 21:37:18 crc kubenswrapper[4726]: I1123 21:37:18.554331 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm5bq" event={"ID":"270843b1-4cd4-4204-a4f5-140276a7b641","Type":"ContainerDied","Data":"c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80"} Nov 23 21:37:18 crc kubenswrapper[4726]: I1123 21:37:18.554361 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm5bq" event={"ID":"270843b1-4cd4-4204-a4f5-140276a7b641","Type":"ContainerStarted","Data":"c8f28a8e5161856f7c170861bc032f5e1b7617fa45844ed0efdb24ac666875b6"} Nov 23 21:37:18 crc kubenswrapper[4726]: I1123 21:37:18.560780 4726 generic.go:334] "Generic (PLEG): container finished" podID="4411be92-5e58-429c-a1ef-cf35c7674665" containerID="a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c" exitCode=0 Nov 23 21:37:18 crc kubenswrapper[4726]: I1123 21:37:18.560818 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47n5x" event={"ID":"4411be92-5e58-429c-a1ef-cf35c7674665","Type":"ContainerDied","Data":"a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c"} Nov 23 21:37:19 crc kubenswrapper[4726]: I1123 21:37:19.574189 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm5bq" event={"ID":"270843b1-4cd4-4204-a4f5-140276a7b641","Type":"ContainerStarted","Data":"d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a"} Nov 23 21:37:19 crc kubenswrapper[4726]: I1123 21:37:19.577374 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47n5x" event={"ID":"4411be92-5e58-429c-a1ef-cf35c7674665","Type":"ContainerStarted","Data":"22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9"} Nov 23 21:37:19 crc kubenswrapper[4726]: I1123 21:37:19.628506 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-47n5x" podStartSLOduration=3.081456628 podStartE2EDuration="5.628491216s" podCreationTimestamp="2025-11-23 21:37:14 +0000 UTC" firstStartedPulling="2025-11-23 21:37:16.43807967 +0000 UTC m=+5344.587120666" lastFinishedPulling="2025-11-23 21:37:18.985114278 +0000 UTC m=+5347.134155254" observedRunningTime="2025-11-23 21:37:19.62078246 +0000 UTC m=+5347.769823416" watchObservedRunningTime="2025-11-23 21:37:19.628491216 +0000 UTC m=+5347.777532172" Nov 23 21:37:21 crc kubenswrapper[4726]: I1123 21:37:21.599650 4726 generic.go:334] "Generic (PLEG): container finished" podID="270843b1-4cd4-4204-a4f5-140276a7b641" containerID="d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a" exitCode=0 Nov 23 21:37:21 crc kubenswrapper[4726]: I1123 21:37:21.599702 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm5bq" event={"ID":"270843b1-4cd4-4204-a4f5-140276a7b641","Type":"ContainerDied","Data":"d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a"} Nov 23 21:37:22 crc kubenswrapper[4726]: I1123 21:37:22.611564 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm5bq" event={"ID":"270843b1-4cd4-4204-a4f5-140276a7b641","Type":"ContainerStarted","Data":"a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7"} Nov 23 21:37:22 crc kubenswrapper[4726]: I1123 21:37:22.635888 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cm5bq" podStartSLOduration=3.195727472 podStartE2EDuration="6.635852777s" podCreationTimestamp="2025-11-23 21:37:16 +0000 UTC" firstStartedPulling="2025-11-23 21:37:18.557391327 +0000 UTC m=+5346.706432283" lastFinishedPulling="2025-11-23 21:37:21.997516622 +0000 UTC m=+5350.146557588" observedRunningTime="2025-11-23 21:37:22.628676097 +0000 UTC m=+5350.777717043" watchObservedRunningTime="2025-11-23 21:37:22.635852777 +0000 UTC m=+5350.784893733" Nov 23 21:37:24 crc kubenswrapper[4726]: I1123 21:37:24.881675 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:24 crc kubenswrapper[4726]: I1123 21:37:24.882244 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:24 crc kubenswrapper[4726]: I1123 21:37:24.945077 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:25 crc kubenswrapper[4726]: I1123 21:37:25.719738 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:26 crc kubenswrapper[4726]: I1123 21:37:26.321499 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47n5x"] Nov 23 21:37:27 crc kubenswrapper[4726]: I1123 21:37:27.289113 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:27 crc kubenswrapper[4726]: I1123 21:37:27.289435 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:27 crc kubenswrapper[4726]: I1123 21:37:27.350054 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:27 crc kubenswrapper[4726]: I1123 21:37:27.661147 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-47n5x" podUID="4411be92-5e58-429c-a1ef-cf35c7674665" containerName="registry-server" containerID="cri-o://22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9" gracePeriod=2 Nov 23 21:37:27 crc kubenswrapper[4726]: I1123 21:37:27.738654 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.117562 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.248479 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-utilities\") pod \"4411be92-5e58-429c-a1ef-cf35c7674665\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.248753 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7qv5\" (UniqueName: \"kubernetes.io/projected/4411be92-5e58-429c-a1ef-cf35c7674665-kube-api-access-q7qv5\") pod \"4411be92-5e58-429c-a1ef-cf35c7674665\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.248780 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-catalog-content\") pod \"4411be92-5e58-429c-a1ef-cf35c7674665\" (UID: \"4411be92-5e58-429c-a1ef-cf35c7674665\") " Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.249253 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-utilities" (OuterVolumeSpecName: "utilities") pod "4411be92-5e58-429c-a1ef-cf35c7674665" (UID: "4411be92-5e58-429c-a1ef-cf35c7674665"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.255480 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4411be92-5e58-429c-a1ef-cf35c7674665-kube-api-access-q7qv5" (OuterVolumeSpecName: "kube-api-access-q7qv5") pod "4411be92-5e58-429c-a1ef-cf35c7674665" (UID: "4411be92-5e58-429c-a1ef-cf35c7674665"). InnerVolumeSpecName "kube-api-access-q7qv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.263783 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4411be92-5e58-429c-a1ef-cf35c7674665" (UID: "4411be92-5e58-429c-a1ef-cf35c7674665"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.350512 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7qv5\" (UniqueName: \"kubernetes.io/projected/4411be92-5e58-429c-a1ef-cf35c7674665-kube-api-access-q7qv5\") on node \"crc\" DevicePath \"\"" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.350544 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.350556 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4411be92-5e58-429c-a1ef-cf35c7674665-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.673492 4726 generic.go:334] "Generic (PLEG): container finished" podID="4411be92-5e58-429c-a1ef-cf35c7674665" containerID="22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9" exitCode=0 Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.673650 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47n5x" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.673673 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47n5x" event={"ID":"4411be92-5e58-429c-a1ef-cf35c7674665","Type":"ContainerDied","Data":"22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9"} Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.674300 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47n5x" event={"ID":"4411be92-5e58-429c-a1ef-cf35c7674665","Type":"ContainerDied","Data":"6b9bb5194e2dc7a3fc7f0f8232130da1291b431690f47e086f604647d15904fc"} Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.674336 4726 scope.go:117] "RemoveContainer" containerID="22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.712379 4726 scope.go:117] "RemoveContainer" containerID="a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.715854 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47n5x"] Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.722916 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-47n5x"] Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.733180 4726 scope.go:117] "RemoveContainer" containerID="f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.807324 4726 scope.go:117] "RemoveContainer" containerID="22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9" Nov 23 21:37:28 crc kubenswrapper[4726]: E1123 21:37:28.807899 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9\": container with ID starting with 22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9 not found: ID does not exist" containerID="22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.807958 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9"} err="failed to get container status \"22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9\": rpc error: code = NotFound desc = could not find container \"22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9\": container with ID starting with 22a40a9950aa21875608cca1c50c1f8be759e81a216977bdfca61b1b52521bd9 not found: ID does not exist" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.807992 4726 scope.go:117] "RemoveContainer" containerID="a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c" Nov 23 21:37:28 crc kubenswrapper[4726]: E1123 21:37:28.808373 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c\": container with ID starting with a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c not found: ID does not exist" containerID="a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.808409 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c"} err="failed to get container status \"a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c\": rpc error: code = NotFound desc = could not find container \"a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c\": container with ID starting with a31e38348081d4390d55e01d9e29acb07c24d56d424e83beedc8e4c80715d38c not found: ID does not exist" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.808430 4726 scope.go:117] "RemoveContainer" containerID="f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97" Nov 23 21:37:28 crc kubenswrapper[4726]: E1123 21:37:28.808667 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97\": container with ID starting with f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97 not found: ID does not exist" containerID="f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97" Nov 23 21:37:28 crc kubenswrapper[4726]: I1123 21:37:28.808701 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97"} err="failed to get container status \"f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97\": rpc error: code = NotFound desc = could not find container \"f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97\": container with ID starting with f944f1542ca863f5964bc91c3e1dc0981fb23afdae3990cbb2364ac411eecc97 not found: ID does not exist" Nov 23 21:37:29 crc kubenswrapper[4726]: I1123 21:37:29.119322 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cm5bq"] Nov 23 21:37:29 crc kubenswrapper[4726]: I1123 21:37:29.688377 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cm5bq" podUID="270843b1-4cd4-4204-a4f5-140276a7b641" containerName="registry-server" containerID="cri-o://a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7" gracePeriod=2 Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.116349 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.189066 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt5mq\" (UniqueName: \"kubernetes.io/projected/270843b1-4cd4-4204-a4f5-140276a7b641-kube-api-access-pt5mq\") pod \"270843b1-4cd4-4204-a4f5-140276a7b641\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.189242 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-catalog-content\") pod \"270843b1-4cd4-4204-a4f5-140276a7b641\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.189353 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-utilities\") pod \"270843b1-4cd4-4204-a4f5-140276a7b641\" (UID: \"270843b1-4cd4-4204-a4f5-140276a7b641\") " Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.190309 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-utilities" (OuterVolumeSpecName: "utilities") pod "270843b1-4cd4-4204-a4f5-140276a7b641" (UID: "270843b1-4cd4-4204-a4f5-140276a7b641"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.196546 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270843b1-4cd4-4204-a4f5-140276a7b641-kube-api-access-pt5mq" (OuterVolumeSpecName: "kube-api-access-pt5mq") pod "270843b1-4cd4-4204-a4f5-140276a7b641" (UID: "270843b1-4cd4-4204-a4f5-140276a7b641"). InnerVolumeSpecName "kube-api-access-pt5mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.250947 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "270843b1-4cd4-4204-a4f5-140276a7b641" (UID: "270843b1-4cd4-4204-a4f5-140276a7b641"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.291418 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt5mq\" (UniqueName: \"kubernetes.io/projected/270843b1-4cd4-4204-a4f5-140276a7b641-kube-api-access-pt5mq\") on node \"crc\" DevicePath \"\"" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.291453 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.291463 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/270843b1-4cd4-4204-a4f5-140276a7b641-utilities\") on node \"crc\" DevicePath \"\"" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.604101 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4411be92-5e58-429c-a1ef-cf35c7674665" path="/var/lib/kubelet/pods/4411be92-5e58-429c-a1ef-cf35c7674665/volumes" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.699168 4726 generic.go:334] "Generic (PLEG): container finished" podID="270843b1-4cd4-4204-a4f5-140276a7b641" containerID="a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7" exitCode=0 Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.699216 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm5bq" event={"ID":"270843b1-4cd4-4204-a4f5-140276a7b641","Type":"ContainerDied","Data":"a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7"} Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.699245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cm5bq" event={"ID":"270843b1-4cd4-4204-a4f5-140276a7b641","Type":"ContainerDied","Data":"c8f28a8e5161856f7c170861bc032f5e1b7617fa45844ed0efdb24ac666875b6"} Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.699266 4726 scope.go:117] "RemoveContainer" containerID="a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.699281 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cm5bq" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.725661 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cm5bq"] Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.736979 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cm5bq"] Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.750968 4726 scope.go:117] "RemoveContainer" containerID="d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.785265 4726 scope.go:117] "RemoveContainer" containerID="c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.811433 4726 scope.go:117] "RemoveContainer" containerID="a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7" Nov 23 21:37:30 crc kubenswrapper[4726]: E1123 21:37:30.811957 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7\": container with ID starting with a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7 not found: ID does not exist" containerID="a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.811988 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7"} err="failed to get container status \"a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7\": rpc error: code = NotFound desc = could not find container \"a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7\": container with ID starting with a66b9f2bd6ff0be57abdbfa3c322f262f748f780707fed66cd34a753a413d8b7 not found: ID does not exist" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.812008 4726 scope.go:117] "RemoveContainer" containerID="d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a" Nov 23 21:37:30 crc kubenswrapper[4726]: E1123 21:37:30.812505 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a\": container with ID starting with d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a not found: ID does not exist" containerID="d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.812527 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a"} err="failed to get container status \"d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a\": rpc error: code = NotFound desc = could not find container \"d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a\": container with ID starting with d73f54545d2806284ea3e83b2dd57650ca0ed1b72e07e718776bd407c771d37a not found: ID does not exist" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.812557 4726 scope.go:117] "RemoveContainer" containerID="c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80" Nov 23 21:37:30 crc kubenswrapper[4726]: E1123 21:37:30.813503 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80\": container with ID starting with c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80 not found: ID does not exist" containerID="c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80" Nov 23 21:37:30 crc kubenswrapper[4726]: I1123 21:37:30.813720 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80"} err="failed to get container status \"c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80\": rpc error: code = NotFound desc = could not find container \"c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80\": container with ID starting with c72e451dae95d4c845d91a7982ea797ef1ab47d37a1ce909aefcdd4313fd0b80 not found: ID does not exist" Nov 23 21:37:32 crc kubenswrapper[4726]: I1123 21:37:32.605911 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="270843b1-4cd4-4204-a4f5-140276a7b641" path="/var/lib/kubelet/pods/270843b1-4cd4-4204-a4f5-140276a7b641/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110677250024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110677251017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110664226016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110664227015461 5ustar corecore